本發明之目的在於提供一字元影像識別方法,該字元影像識別方法包括以下步驟: (a) 匯入一待識別文件,該待識別文件包括多個字元影像; (b) 辨識出至少一字元分割區域,該字元分割區域包括至少一上述字元影像; (c) 將該字元分割區域中的該字元影像轉換為一可編輯字元; (d) 對該可編輯字元進行校對; (e) 辨識出該字元分割區域於該待識別文件中的相對位置,並對該字元分割區域進行標籤;以及 (f) 輸出該可編輯字元、該字元分割區域於該待識別文件中的相對位置、與該字元分割區域所對應的標籤。 如上述之字元影像識別方法,其中於(b)步驟中,該字元分割區域被分類為一需求字元集與一非需求字元集。 如上述之字元影像識別方法,其中於(b)步驟中是藉由一第一類神經網路模型辨識出該字元分割區域。 如上述之字元影像識別方法,其中該第一類神經網路模型於訓練階段時,採用下述的步驟進行訓練: (b1) 針對該待識別文件,獲取一第一數量的同類別的影像樣本; (b2) 將該些影像樣本分成為一訓練集、一測試集、與一驗證集; (b3) 利用該訓練集作為訓練樣本,以生成該第一類神經網路模型的原型; (b4) 利用該測試集驗證該第一類神經網路模型的原型的正確性,若該測試集的驗證通過率小於一第一預設閾值,則進入步驟(b5),若該測試集的驗證通過率不小於該第一預設閾值,則進入步驟(b6); (b5) 再獲取一第二數量的同類別的影像樣本,並重複(b2)~ (b4)的步驟; (b6) 利用該驗證集驗證該第一類神經網路模型的正確性,若該驗證集的驗證通過率小於一第二預設閾值,則進入步驟(b7),若該測試集的驗證通過率不小於該第二預設閾值,則進入步驟(b8); (b7) 再獲取一第三數量的同類別的影像樣本,並重複 (b6)的步驟; (b8) 完成該第一類神經網路模型的訓練。 如上述之字元影像識別方法,其中於(d)步驟中是藉由一第二類神經網路模型對該可編輯字元進行校對。 如上述之字元影像識別方法,其中該第二類神經網路模型於訓練階段時,採用下述的步驟進行訓練: (d1) 針對該待識別文件中的該字元分割區域,獲取一第四數量的該字元分割區域的可編輯字元樣本; (d2) 將該些可編輯字元樣本分成為一訓練集、一測試集、與一驗證集; (d3) 利用該訓練集作為訓練樣本,以生成該第二類神經網路模型的原型; (d4) 利用該測試集驗證該第二類神經網路模型的正確性,若該測試集的驗證通過率小於一第四預設閾值,則進入步驟(d5),若該測試集的驗證通過率不小於該第四預設閾值,則進入步驟(d6); (d5) 再獲取一第五數量的同類別的可編輯字元樣本,並重複(d2)~ (d4)的步驟; (d6) 利用該驗證集驗證該字元提取模型的正確性,若該驗證集的驗證通過率小於一第五預設閾值,則進入步驟(d7),若該測試集的驗證通過率不小於該第五預設閾值,則進入步驟(d8); (d7) 再獲取一第六數量的同類別的可編輯字元樣本,並重複(d2)~ (d6)的步驟; (d8) 完成該第二類神經網路模型的訓練。 如上述之字元影像識別方法,其中於(e)步驟中是藉由一第三類神經網路模型對辨識出該字元分割區域於該待識別文件中的相對位置,並對該字元分割區域進行標籤。 如上述之字元影像識別方法,其中該第三類神經網路模型於訓練階段時,採用下述的步驟進行訓練: (e1) 針對該待識別文件中的該字元分割區域,獲取一第七數量的該字元分割區域的標籤與在該待識別文件中的相對位置; (e2) 將該些標籤樣本分成為一訓練集、一測試集、與一驗證集; (e3) 利用該訓練集作為訓練樣本,以生成該第三類神經網路模型的原型; (e4) 利用該測試集驗證該第三類神經網路模型的正確性,若該測試集的驗證通過率小於一第七預設閾值,則進入步驟(e5),若該測試集的驗證通過率不小於該第七預設閾值,則進入步驟(e6); (e5) 再獲取一第八數量的同類別的標籤樣本,並重複(e2)~ (e4)的步驟; (e6) 利用該驗證集驗證該第三類神經網路模型的正確性,若該驗證集的驗證通過率小於一第八預設閾值,則進入步驟(e7),若該測試集的驗證通過率不小於該第八預設閾值,則進入步驟(e8); (e7) 再獲取一第九數量的同類別的標籤樣本,並重複(e2)~ (e6)的步驟; (e8) 完成該第三類神經網路模型的訓練。 本發明之一種字元影像識別系統,用以識別一待識別文件中的字元,該待識別文件包括多個字元影像,該字元影像識別系統包括一字元分割區域識別模組、一文意分析模組、一字元分割區域標籤模組及一輸出模組。其中,一字元分割區域識別模組,辨識出至少一字元分割區域,該字元分割區域包括至少一上述字元影像。此外,一文意分析模組,通信連結該字元分割區域識別模組,該文意分析模組將該字元分割區域中的該字元影像轉換為一可編輯字元,並對該可編輯字元進行校對。而一字元分割區域標籤模組,該字元分割區域標籤模組識別出該字元分割區域於該待識別文件中的相對位置,並對該字元分割區域進行標籤,及一輸出模組,輸出該可編輯字元、該字元分割區域於該待識別文件中的相對位置、與該字元分割區域所對應的標籤。 如上述之字元影像識別系統,其中該字元影像識別系統還包括一伺服端與一客戶端,該客戶端具有一顯示螢幕,其中該字元分割區域識別模組、該文意分析模組、該字元分割區域標籤模組、與該輸出模組是設置於該伺服端,該輸出模組輸出該可編輯字元至該客戶端並顯示於該顯示螢幕上。 為讓本之上述特徵和優點能更明顯易懂,下文特舉較佳實施例,並配合所附圖式,作詳細說明如下。The object of the present invention is to provide a character image recognition method, which includes the following steps: (a) importing a document to be recognized, the document to be recognized includes a plurality of character images; (b) recognizing at least A character segmented area, the character segmented area including at least one of the above-mentioned character images; (c) the character image in the character segmented area is converted into an editable character; (d) the editable character (E) Identify the relative position of the character segmentation area in the document to be recognized, and label the character segmentation area; and (f) Output the editable character and the character segmentation area The relative position in the document to be recognized and the label corresponding to the character segmentation area. As in the above-mentioned character image recognition method, in step (b), the character segmentation area is classified into a required character set and an unrequired character set. In the above-mentioned character image recognition method, in step (b), the character segmentation area is recognized by a first-type neural network model. Such as the above-mentioned character image recognition method, wherein the first type of neural network model is trained in the following steps during the training phase: (b1) For the document to be recognized, a first number of images of the same type are obtained Samples; (b2) Divide the image samples into a training set, a test set, and a validation set; (b3) Use the training set as a training sample to generate the prototype of the first type of neural network model; ( b4) Use the test set to verify the correctness of the prototype of the first type of neural network model. If the verification pass rate of the test set is less than a first preset threshold, proceed to step (b5), if the test set is verified If the pass rate is not less than the first preset threshold, proceed to step (b6); (b5) Obtain a second number of image samples of the same category, and repeat steps (b2) ~ (b4); (b6) Use The verification set verifies the correctness of the first type of neural network model. If the verification pass rate of the verification set is less than a second preset threshold, then step (b7) is entered. If the verification pass rate of the test set is not less than the Go to step (b8) for the second preset threshold; (b7) Obtain a third number of image samples of the same type, and repeat the step (b6); (b8) Complete the first type of neural network model training. In the above-mentioned character image recognition method, in step (d), the editable character is proofread by a second-type neural network model. Such as the above-mentioned character image recognition method, wherein the second type of neural network model is trained in the following steps during the training stage: (d1) For the character segmentation area in the document to be recognized, a first Four editable character samples of the character segmentation area; (d2) Divide the editable character samples into a training set, a test set, and a verification set; (d3) Use the training set as training Sample to generate the prototype of the second type neural network model; (d4) use the test set to verify the correctness of the second type neural network model, if the verification pass rate of the test set is less than a fourth preset threshold , Then go to step (d5), if the verification pass rate of the test set is not less than the fourth preset threshold, go to step (d6); (d5) get a fifth number of editable character samples of the same category , And repeat the steps (d2)~(d4); (d6) Use the verification set to verify the correctness of the character extraction model. If the verification pass rate of the verification set is less than a fifth preset threshold, then proceed to step ( d7), if the verification pass rate of the test set is not less than the fifth preset threshold, proceed to step (d8); (d7) obtain a sixth number of editable character samples of the same category, and repeat (d2 )~ (d6) steps; (d8) Complete the training of the second type of neural network model. As in the above-mentioned character image recognition method, in step (e), a third type neural network model is used to identify the relative position of the character segmentation area in the document to be recognized, and the character Separate the area for labeling. As in the above-mentioned character image recognition method, the third type of neural network model is trained in the following steps during the training phase: (e1) For the character segmentation area in the document to be recognized, a first The relative positions of the labels of the seven-quantity character segmentation area in the document to be recognized; (e2) Divide the label samples into a training set, a test set, and a verification set; (e3) Use the training Set as training samples to generate the prototype of the third type of neural network model; (e4) Use the test set to verify the correctness of the third type of neural network model, if the verification pass rate of the test set is less than a seventh If the preset threshold is set, then go to step (e5), if the verification pass rate of the test set is not less than the seventh preset threshold, then go to step (e6); (e5) get an eighth number of label samples of the same category , And repeat the steps (e2)~(e4); (e6) Use the verification set to verify the correctness of the third type of neural network model, if the verification pass rate of the verification set is less than an eighth preset threshold, then Go to step (e7), if the verification pass rate of the test set is not less than the eighth preset threshold, go to step (e8); (e7) get a ninth number of label samples of the same category, and repeat (e2 )~ (e6) steps; (e8) Complete the training of the third type of neural network model. A character image recognition system of the present invention is used for recognizing characters in a document to be recognized. The document to be recognized includes a plurality of character images. The character image recognition system includes a character segmentation area recognition module, a text Meaning analysis module, a character segmentation area label module and an output module. Wherein, a character segmentation area recognition module recognizes at least one character segmentation area, and the character segmentation area includes at least one character image. In addition, a context analysis module is communicatively connected to the character segmentation area recognition module, and the context analysis module converts the character image in the character segmentation area into an editable character, and the editable Characters are proofread. And a character segmentation area labeling module, the character segmentation area labeling module recognizes the relative position of the character segmentation area in the document to be recognized, and labels the character segmentation area, and an output module , Output the editable character, the relative position of the character segmentation area in the file to be recognized, and the label corresponding to the character segmentation area. Such as the above-mentioned character image recognition system, wherein the character image recognition system further includes a server and a client, the client has a display screen, wherein the character segmentation area recognition module and the context analysis module , The character segmentation area label module, and the output module are arranged on the server, and the output module outputs the editable character to the client and displays it on the display screen. In order to make the above-mentioned features and advantages of the present invention more obvious and easy to understand, the following is a detailed description of preferred embodiments in conjunction with the accompanying drawings.
參照本文闡述的詳細內容和附圖說明是最好理解本發明。下面參照附圖會討論各種實施例。然而,本領域技術人員將容易理解,這裡關於附圖給出的詳細描述僅僅是為了解釋的目的,因為這些方法和系統可超出所描述的實施例。例如,所給出的教導和特定應用的需求可能產生多種可選的和合適的方法來實現在此描述的任何細節的功能。因此,任何方法可延伸超出所描述和示出的以下實施例中的特定實施選擇範圍。 在說明書及後續的申請專利範圍當中使用了某些詞彙來指稱特定的元件。所屬領域中具有通常知識者應可理解,硬體製造商可能會用不同的名詞來稱呼同樣的元件。本說明書及後續的申請專利範圍並不以名稱的差異來作為區分元件的方式,而是以元件在功能上的差異來作為區分的準則。在通篇說明書及後續的請求項當中所提及的「包含」係為一開放式的用語,故應解釋成「包含但不限定於」。另外,「耦接」一詞在此係包含任何直接及間接的電氣連接手段。因此,若文中描述一第一裝置耦接於一第二裝置,則代表該第一裝置可直接電氣連接於該第二裝置,或透過其他裝置或連接手段間接地電氣連接至該第二裝置。 請參閱圖1,圖1所繪示為本發明之字元影像識別系統的實施例。字元影像識別系統100包括一字元分割區域識別模組110、一文意分析模組120、一字元分割區域標籤模組130與一輸出模組140,其中字元影像識別系統100還電性連接到一影像輸入裝置10,此影像輸入裝置10例如為一掃描裝置或一數位相機。藉由此影像輸入裝置10,可將一待識別文件(如圖2A)匯入到字元影像識別系統100中。在本實施例中,字元分割區域識別模組110、文意分析模組120、字元分割區域標籤模組130與輸出模組140是設置於伺服端102,該伺服端102例如是由一台或多台伺服器所組成。另外,輸出模組140是電性連接到一客戶端104,客戶端104可為一具有顯示螢幕104a的電子裝置,此電子裝置例如為個人電腦、筆記型電腦、或智慧型手機。 另外,也請參照圖2A,圖2A所繪示為待識別文件的其中一實施例,在本實施例中待識別文件為診斷證明書。從圖2A可知,此待識別文件包括多個字元,而當待識別文件的影像被影像輸入裝置10捕捉後,待識別文件上的字元當然也是以影像的方式存在的,也就是說由影像輸入裝置10匯入到字元影像識別系統100的待識別文件上的字元是無法編輯的,以下將這些字元稱為字元影像。 此外,請同時參照圖3,圖3所繪示為本發明之字元影像識別方法的實施例。首先,實施步驟S210,匯入如圖2A的待識別文件,其詳細流程已如上文所述,在此不再贅述。接著,實施步驟S220,辨識出待識別文件中的字元分割區域81。在圖2B中,字元分割區域81是由虛線所框出來的區域,字元分割區域81例如是由字元分割區域識別模組110識別出來。由圖2B可清楚得知,字元分割區域81是將待識別文件上的字元影像選取出來。在較佳的實施例中,實施步驟S220後還會實施步驟S222,將字元分割區域81分成一需求字元集與一非需求字元集。需求字元集是指在後續的處理中,需要輸出的字元的集合,這些字元例如為圖2B中的元件符號81a所標示出來的區域。非需求字元集是指在後續的處理中,不需要輸出的字元的集合,這些字元例如為圖2B中的元件符號81b所標示出來的區域。更詳細的舉例,在圖2B中,印章『以下空白』由於對於之後的資料處理可能無關輕重,故被歸類為非需求字元集。 之後,實施步驟S230,藉由文意分析模組120將字元分割區域81中的字元影像轉換為可編輯字元。也就是說,原本由影像輸入裝置10所匯入的待識別文件的影像,其上的字元影像是無法編輯的;然而文意分析模組120可將這些字元影像轉換為可編輯字元,其例如是採用OCR(Optical Character Recognition,光學字元識別)的技術。然而,若單純採用OCR的技術,在待識別文件上的字元影像模糊不清或是被髒污附著時,便可能發生判別錯誤的情形。舉例來說,原本『雙和醫院』因為被髒污附著而產生如圖4上方所示的字元,若單純只採用OCR的技術,便可能將『雙和醫院』辨識為『雙利醫院』(如圖4中央所示)。然而,在本實施例中,文意分析模組120可執行步驟S240,對所轉換出的可編輯字元進行校正,例如將『雙利醫院』校正成『雙和醫院』(如圖4下方所示)。 之後,實施步驟S250,可藉由字元分割區域標籤模組130辨識出各個字元分割區域81於待識別文件中的相對位置,並對字元分割區域81進行標籤。舉例來說,在圖2B的待識別文件中,字元分割區域標籤模組130可辨識出『姓名』這個欄位在待識別文件中的相對位置,同時也『姓名』這個欄位一個叫做『姓名』的標籤。 再來,實施步驟S260,藉由輸出模組140輸出上述的可編輯字元、字元分割區域81於待識別文件中的相對位置、與字元分割區域81所對應的標籤至客戶端104,這樣便可在客戶端104的顯示螢幕104a上顯示待識別文件上的字元。值得注意的是,由於輸出模組140還會輸出字元分割區域81於待識別文件中的相對位置,故在客戶端104的顯示螢幕104a上還會重現各字元於待識別文件上的位置。也就是說,若待識別文件為表格形式的文件(如圖2A所示),則在經過本發明之字元影像識別系統100的處理後,不但能呈現可編輯字元還能重現這些可編輯字元於表格中的位置,這是一般OCR軟體所無法達到的。 此外,由於字元分割區域81的標籤也會被輸出,故客戶端104更可對所接受的資料進行更進一步的處理,例如製作成Excel表或由資料庫管理系統對待識別文件中的資料進行整理和歸類。舉例來說,若藉由一般的OCR軟體來擷取診斷證明書的資料,即使準確的辨別各字元,使用者還是需用人工的方式來設定病人與病名的對應關係。然而,藉由字元影像識別系統100,便可自動地建立起病人與病名的對應關係,從而減少人力成本。 綜上,相較於一般的OCR軟體,本發明之字元影像識別系統100具有對辨識錯誤的字元進行校正、重現字元於待識別文件中的相對位置、與對字元分割區域81進行標籤以利於後續資料處理等優點。 在上述的實施例中,字元分割區域識別模組110例如是藉由一第一類神經網路模型112辨識出待識別文件中的字元分割區域81,第一類神經網路模型112例如是藉由深度學習的方式進行訓練,以期提高第一類神經網路模型112的精度,以下將介紹第一類神經網路模型112的訓練方式,還請參照圖5,圖5所繪示為第一類神經網路模型於訓練階段的流程圖。首先,實施步驟S310,針對某類別的待識別文件(例如圖2A所示的診斷證明書),取得一定數量N1的同類別的影像樣本(為了區分方便,在本文其他處又可將N1稱為第一數量)。接著,實施步驟S320,將收集到的這些影像樣本分成為一訓練集、一測試集、與一驗證集。接著,實施步驟S330,將訓練集作為一神經網路模型的訓練樣本,該神經網路模型完成訓練後便生成該第一類神經網路模型112的原型。 之後,實施步驟S340,利用測試集驗證第一類神經網路模型112的原型的正確性,若驗證通過率小於一第一預設閾值(例如90%),則實施步驟S350,若否則實施步驟S360。在此,所謂的通過是指第一類神經網路模型112可以正確辨識出字元分割區域81,在較佳的實施例中還可進一步將字元分割區域81分類為需求字元集與非需求字元集。 在步驟S350中,需再獲取再一定數量N2的同類別的影像樣本(為了區分方便,在本文其他處又可將N2稱為第二數量),並重複步驟S320~步驟S340。在步驟S360中,利用驗證集驗證第一類神經網路模型112的正確性,若驗證通過率小於一第二預設閾值(例如95%),則實施步驟S370,若否則實施步驟S380。 在步驟S370中,需再獲取又一定數量N3的同類別的影像樣本(為了區分方便,在本文其他處又可將N3稱為第三數量),並重複步驟S360。在步驟S380中,則完成第一類神經網路模型112的訓練,也就是說可以將第一類神經網路模型112投入實用中。 在上述的實施例中,文意分析模組120例如是藉由一第二類神經網路模型122對可編輯字元進行校正,例如如圖4所示將『雙利醫院』校正成『雙和醫院』。第二類神經網路模型122例如是藉由深度學習的方式進行訓練,以期提高第二類神經網路模型122的精度,以下將介紹第二類神經網路模型122的訓練方式,還請參照圖6,圖6所繪示為第二類神經網路模型於訓練階段的流程圖。首先,實施步驟S410,針對待識別文件中的某一字元分割區域81,獲取一定數量N4的該字元分割區域的可編輯字元樣本(為了區分方便,在本文其他處又可將N4稱為第四數量)。舉例來說,圖2B中的元件符號81a所標示出來的區域是指醫院名稱。在步驟S410中,就是輸入數量為第四數量N4的醫院名稱。 接著,實施步驟S420,將收集到的這些可編輯字元樣本分成為一訓練集、一測試集、與一驗證集。接著,實施步驟S430,將訓練集作為一神經網路模型的訓練樣本,該神經網路模型完成訓練後便生成第二類神經網路模型122的原型。之後,實施步驟S440,利用測試集驗證第二類神經網路模型122的原型的正確性,若驗證通過率小於一第一預設閾值(例如90%),則實施步驟S450,若否則實施步驟S460。在此,所謂的通過是指第二類神經網路模型122可以正確校正可編輯字元的錯誤。 在步驟S450中,需再獲取再一定數量N5的同類別的可編輯字元樣本(為了區分方便,在本文其他處又可將N5稱為第五數量),並重複步驟S420~步驟S440。在步驟S460中,利用驗證集驗證第二類神經網路模型122的正確性,若驗證通過率小於一第二預設閾值(例如95%),則實施步驟S470,若否則實施步驟S480。 在步驟S470中,需再獲取又一定數量N6的同類別的可編輯字元樣本(為了區分方便,在本文其他處又可將N6稱為第六數量),並重複步驟S460。在步驟S480中,則完成第二類神經網路模型122的訓練,也就是說可以將第二類神經網路模型122投入實用中。 在上述的實施例中,字元分割區域標籤模組130例如是藉由一第三類神經網路模型132辨識出各個字元分割區域81於待識別文件中的相對位置,並對字元分割區域81進行標籤。第三類神經網路模型132例如是藉由深度學習的方式進行訓練,以期提高第三類神經網路模型132的精度,以下將介紹第三類神經網路模型132的訓練方式,還請參照圖7,圖7所繪示為第三類神經網路模型於訓練階段的流程圖。首先,實施步驟S510,針對待識別文件中的字元分割區域81,獲取一第七數量N7的該字元分割區域81的標籤與在該待識別文件中的相對位置。字元分割區域81的標籤例如是由人工先進行標註後再輸入一神經網路模型,例如可以聘請多位資料標註師標註醫院名稱在待識別文件的哪邊,病人的姓名又在待識別文件的何處。關於如何標註,可以參考本案申請人的另外一篇專利申請案(申請號:107140893)。 接著,實施步驟S520,將收集到的這些字元分割區域81的標籤的樣本分成為一訓練集、一測試集、與一驗證集。接著,實施步驟S530,將訓練集作為一神經網路模型的訓練樣本,該神經網路模型完成訓練後便生成第三類神經網路模型132的原型。之後,實施步驟S540,利用測試集驗證第三類神經網路模型132的原型的正確性,若驗證通過率小於一第一預設閾值(例如90%),則實施步驟S550,若否則實施步驟S560。在此,所謂的通過是指第三類神經網路模型132可以正確辨識出各個字元分割區域81於待識別文件中的相對位置,並對字元分割區域81進行標籤。 在步驟S550中,需再獲取再一定數量N8的同類別的標籤樣本(為了區分方便,在本文其他處又可將N8稱為第八數量),並重複步驟S520~步驟S540。在步驟S560中,利用驗證集驗證第三類神經網路模型132的正確性,若驗證通過率小於一第二預設閾值(例如95%),則實施步驟S570,若否則實施步驟S580。 在步驟S570中,需再獲取又一定數量N9的同類別的標籤樣本(為了區分方便,在本文其他處又可將N9稱為第九數量),並重複步驟S560。在步驟S580中,則完成第三類神經網路模型132的訓練,也就是說可以將第三類神經網路模型132投入實用中。 在上述的實施例中,不管是第一類神經網路模型112、第二類神經網路模型122、或第三類神經網路模型132,於訓練時都是將樣本分成訓練集、測試集、與驗證集,先由訓練集訓練後,再由測試集進行測試,若通過則再由驗證集進行驗證。相較於習知的神經網路模型在訓練時,只將樣本分成訓練集與測試集,本案之第一類神經網路模型112、第二類神經網路模型122、與第三類神經網路模型132可獲得更高的正確率。此外,在上述的實施例中,第一類神經網路模型112、第二類神經網路模型122、或第三類神經網路模型132可以為可為遞歸神經網路(Recurrent Neural Network)、長短期記憶模型(Long Short-Term Memory)或是卷積神經網路(Convolutional Neural Network),請注意,此僅為本發明的實施例,並非本發明的限制條件。 雖然本發明已以較佳實施例揭露如上,然其並非用以限定本發明,任何所屬技術領域中具有通常知識者,在不脫離本發明之精神和範圍內,當可作些許之更動與潤飾,因此本發明之保護範圍當視後附之申請專利範圍所界定者為準。The present invention is best understood with reference to the detailed content set forth herein and the description of the drawings. Various embodiments will be discussed below with reference to the drawings. However, those skilled in the art will easily understand that the detailed description given here with respect to the drawings is only for the purpose of explanation, because these methods and systems may go beyond the described embodiments. For example, the given teachings and the requirements of specific applications may produce a variety of alternative and suitable methods to achieve any detailed functions described herein. Therefore, any method can extend beyond the specific implementation options described and illustrated in the following embodiments. Certain words are used in the specification and subsequent patent applications to refer to specific components. Those with general knowledge in the field should understand that hardware manufacturers may use different terms to refer to the same components. The scope of this specification and subsequent patent applications does not use differences in names as a way of distinguishing elements, but uses differences in functions as a criterion for distinguishing. The "include" mentioned in the entire specification and subsequent requests is an open term, so it should be interpreted as "include but not limited to." In addition, the term "coupling" here includes any direct and indirect electrical connection means. Therefore, if it is described that a first device is coupled to a second device, it means that the first device can be directly electrically connected to the second device, or indirectly electrically connected to the second device through other devices or connecting means. Please refer to FIG. 1. FIG. 1 illustrates an embodiment of the character image recognition system of the present invention. The character image recognition system 100 includes a character segmentation area recognition module 110, a context analysis module 120, a character segmentation area labeling module 130, and an output module 140. The character image recognition system 100 is also electronic Connected to an image input device 10, the image input device 10 is, for example, a scanning device or a digital camera. With this image input device 10, a document to be recognized (as shown in FIG. 2A) can be imported into the character image recognition system 100. In this embodiment, the character segmentation area recognition module 110, the context analysis module 120, the character segmentation area labeling module 130, and the output module 140 are provided on the server 102, and the server 102 is composed of, for example, a It is composed of one or more servers. In addition, the output module 140 is electrically connected to a client 104, which may be an electronic device with a display screen 104a, such as a personal computer, a notebook computer, or a smart phone. In addition, please also refer to FIG. 2A. FIG. 2A shows an embodiment of the document to be identified. In this embodiment, the document to be identified is a diagnosis certificate. It can be seen from FIG. 2A that the document to be recognized includes a plurality of characters, and when the image of the document to be recognized is captured by the image input device 10, the characters on the document to be recognized certainly exist in the form of images, that is to say by The characters on the to-be-recognized document imported by the image input device 10 into the character image recognition system 100 cannot be edited, and these characters are referred to as character images below. In addition, please refer to FIG. 3 at the same time. FIG. 3 illustrates an embodiment of the character image recognition method of the present invention. First, step S210 is implemented to import the file to be identified as shown in FIG. 2A. The detailed flow is as described above, and will not be repeated here. Next, step S220 is implemented to identify the character segmentation area 81 in the document to be identified. In FIG. 2B, the character segmentation area 81 is a region framed by a dotted line, and the character segmentation area 81 is recognized by the character segmentation area recognition module 110, for example. It can be clearly seen from FIG. 2B that the character segmentation area 81 selects the character image on the document to be recognized. In a preferred embodiment, after step S220 is performed, step S222 is also performed to divide the character segmentation area 81 into a required character set and an unrequired character set. The required character set refers to the set of characters that need to be output in the subsequent processing, and these characters are, for example, the area marked by the component symbol 81a in FIG. 2B. The non-required character set refers to the set of characters that do not need to be output in the subsequent processing. These characters are, for example, the area marked by the component symbol 81b in FIG. 2B. For a more detailed example, in FIG. 2B, the seal "blank below" is classified as a non-required character set because it may be irrelevant for subsequent data processing. After that, step S230 is implemented to convert the character image in the character segmentation area 81 into editable characters by the context analysis module 120. In other words, the character images on the image of the document to be recognized originally imported by the image input device 10 cannot be edited; however, the textual analysis module 120 can convert these character images into editable characters For example, it uses OCR (Optical Character Recognition, optical character recognition) technology. However, if the OCR technology is used alone, when the character image on the document to be recognized is blurred or dirty, the judgment error may occur. For example, the original "Shuanghe Hospital" was created with the characters shown in the upper part of Figure 4 because of the dirt attached. If only OCR technology is used, "Shuanghe Hospital" may be identified as "Shuangli Hospital". (As shown in the center of Figure 4). However, in this embodiment, the textual analysis module 120 can perform step S240 to correct the converted editable characters, for example, correcting "Shuangli Hospital" to "Shuanghe Hospital" (Figure 4 below) Shown). After that, step S250 is implemented to identify the relative position of each character segment area 81 in the document to be recognized by the character segment area labeling module 130, and label the character segment area 81. For example, in the document to be recognized in FIG. 2B, the character segmentation area label module 130 can recognize the relative position of the field "name" in the document to be recognized, and also the field of "name" is called " "Name" tag. Then, step S260 is implemented to output the above-mentioned editable characters, the relative position of the character segment area 81 in the document to be recognized, and the label corresponding to the character segment area 81 to the client 104 through the output module 140, In this way, the characters on the document to be recognized can be displayed on the display screen 104a of the client 104. It is worth noting that since the output module 140 also outputs the relative position of the character segmented area 81 in the document to be recognized, the display screen 104a of the client 104 will also reproduce the characters on the document to be recognized. position. That is to say, if the document to be recognized is a table format document (as shown in FIG. 2A), after being processed by the character image recognition system 100 of the present invention, not only editable characters can be presented but also these characters can be reproduced. The position of the edit character in the table is beyond the reach of ordinary OCR software. In addition, since the label of the character segmented area 81 will also be output, the client 104 can further process the received data, such as making it into an Excel table or processing the data in the document to be identified by the database management system. Organize and categorize. For example, if general OCR software is used to retrieve the data of the diagnosis certificate, even if the characters are accurately identified, the user still needs to manually set the correspondence between the patient and the disease name. However, with the character image recognition system 100, the correspondence between the patient and the disease name can be automatically established, thereby reducing labor costs. To sum up, compared with general OCR software, the character image recognition system 100 of the present invention has the functions of correcting incorrectly recognized characters, reproducing the relative position of the characters in the document to be recognized, and dividing the character area 81 Labeling to facilitate subsequent data processing and other advantages. In the above-mentioned embodiment, the character segmentation area recognition module 110 uses a first type neural network model 112 to identify the character segmentation area 81 in the document to be recognized, for example, It is trained by deep learning in order to improve the accuracy of the first type of neural network model 112. The training method of the first type of neural network model 112 will be introduced below. Please also refer to FIG. 5, which is shown as The flow chart of the first type of neural network model in the training phase. First, step S310 is implemented to obtain a certain number of N1 image samples of the same type for a certain type of document to be identified (such as the diagnosis certificate shown in Figure 2A) (for the convenience of distinction, N1 can be referred to as The first quantity). Then, step S320 is implemented to divide the collected image samples into a training set, a test set, and a verification set. Next, step S330 is implemented to use the training set as a training sample of a neural network model. After the neural network model is trained, a prototype of the first type neural network model 112 is generated. After that, step S340 is implemented to verify the correctness of the prototype of the first-type neural network model 112 using the test set. If the verification pass rate is less than a first preset threshold (for example, 90%), step S350 is implemented, if otherwise, step S350 is implemented. S360. Here, the so-called pass means that the first type of neural network model 112 can correctly identify the character segmentation area 81. In a preferred embodiment, the character segmentation area 81 can be further classified into a required character set and a non- Required character set. In step S350, another certain number of N2 image samples of the same category need to be obtained (for the convenience of distinction, N2 may be referred to as the second number elsewhere in this document), and steps S320 to S340 are repeated. In step S360, the verification set is used to verify the correctness of the first-type neural network model 112. If the verification pass rate is less than a second preset threshold (for example, 95%), step S370 is performed, and if not, step S380 is performed. In step S370, another certain number of N3 image samples of the same category need to be obtained (for the convenience of distinction, N3 may be referred to as the third number elsewhere in this document), and step S360 is repeated. In step S380, the training of the first type of neural network model 112 is completed, that is, the first type of neural network model 112 can be put into practice. In the above-mentioned embodiment, the textual analysis module 120 uses a second-type neural network model 122 to correct editable characters. For example, as shown in FIG. 4, "Shuangli Hospital" is corrected to "Double And the hospital. The second type of neural network model 122 is trained by, for example, deep learning, in order to improve the accuracy of the second type of neural network model 122. The following will introduce the training method of the second type of neural network model 122, please refer to Figure 6, Figure 6 shows a flowchart of the second type of neural network model in the training phase. First, step S410 is performed to obtain a certain number of N4 editable character samples in the character segmentation area 81 for a character segmentation area 81 in the document to be recognized (for the convenience of distinction, N4 can be referred to elsewhere in this article Is the fourth quantity). For example, the area marked by the symbol 81a in FIG. 2B refers to the name of the hospital. In step S410, the hospital name of the fourth number N4 is input. Then, step S420 is implemented to divide the collected editable character samples into a training set, a test set, and a verification set. Next, step S430 is implemented to use the training set as a training sample of a neural network model, and the neural network model generates a prototype of the second type of neural network model 122 after completing the training. After that, step S440 is implemented to verify the correctness of the prototype of the second-type neural network model 122 using the test set. If the verification pass rate is less than a first preset threshold (for example, 90%), step S450 is implemented, if otherwise, step S450 is implemented. S460. Here, the so-called pass means that the second type neural network model 122 can correctly correct the errors of editable characters. In step S450, a certain number of N5 editable character samples of the same type need to be obtained (for the convenience of distinction, N5 may be referred to as the fifth number elsewhere in this document), and steps S420 to S440 are repeated. In step S460, the verification set is used to verify the correctness of the second-type neural network model 122. If the verification pass rate is less than a second preset threshold (for example, 95%), step S470 is implemented, and if not, step S480 is implemented. In step S470, another certain number of N6 editable character samples of the same category need to be obtained (for distinguishing convenience, N6 may be referred to as the sixth number elsewhere in this document), and step S460 is repeated. In step S480, the training of the second type of neural network model 122 is completed, that is, the second type of neural network model 122 can be put into practical use. In the above-mentioned embodiment, the character segmentation area labeling module 130, for example, uses a third-type neural network model 132 to identify the relative position of each character segmentation area 81 in the document to be recognized, and segment the characters Area 81 is labeled. The third type of neural network model 132 is trained by deep learning, for example, in order to improve the accuracy of the third type of neural network model 132. The following will introduce the training method of the third type of neural network model 132, please refer to Figure 7, Figure 7 shows a flowchart of the third type of neural network model in the training phase. First, step S510 is implemented to obtain a seventh number N7 of the label of the character segmented area 81 and the relative position in the file to be recognized for the character segmented area 81 in the file to be recognized. The label of the character segmentation area 81 is, for example, manually annotated and then input into a neural network model. For example, multiple data annotators can be hired to mark where the hospital name is on the document to be identified, and the patient's name is in the document to be identified Where. For how to label, you can refer to another patent application (application number: 107140893) of the applicant in this case. Next, step S520 is implemented to divide the collected label samples of the character segmentation area 81 into a training set, a test set, and a verification set. Then, step S530 is implemented, and the training set is used as a training sample of a neural network model. After the neural network model is trained, a prototype of the third type of neural network model 132 is generated. After that, step S540 is implemented to verify the correctness of the prototype of the third-type neural network model 132 using the test set. If the verification pass rate is less than a first preset threshold (for example, 90%), step S550 is implemented, if otherwise, step S550 is implemented. S560. Here, the so-called pass means that the third-type neural network model 132 can correctly recognize the relative position of each character segmentation area 81 in the document to be recognized, and label the character segmentation area 81. In step S550, it is necessary to obtain a certain number of N8 label samples of the same category (for distinguishing convenience, N8 may be referred to as the eighth number elsewhere in this document), and steps S520 to S540 are repeated. In step S560, the verification set is used to verify the correctness of the third-type neural network model 132. If the verification pass rate is less than a second preset threshold (for example, 95%), step S570 is performed, and if not, step S580 is performed. In step S570, another certain number of N9 label samples of the same category need to be obtained (for distinguishing convenience, N9 may be referred to as the ninth number elsewhere in this document), and step S560 is repeated. In step S580, the training of the third type of neural network model 132 is completed, that is, the third type of neural network model 132 can be put into practical use. In the above-mentioned embodiment, whether it is the first type of neural network model 112, the second type of neural network model 122, or the third type of neural network model 132, the samples are divided into training sets and test sets during training. , And the validation set, first trained by the training set, and then tested by the test set, and if it passes, it will be verified by the validation set. Compared with the conventional neural network model during training, only the sample is divided into training set and test set. In this case, the first type of neural network model 112, the second type of neural network model 122, and the third type of neural network The road model 132 can obtain a higher accuracy rate. In addition, in the above-mentioned embodiment, the first type of neural network model 112, the second type of neural network model 122, or the third type of neural network model 132 may be a recurrent neural network (Recurrent Neural Network), Long short-term memory model (Long Short-Term Memory) or Convolutional Neural Network (Convolutional Neural Network), please note that this is only an embodiment of the present invention, not a limitation of the present invention. Although the present invention has been disclosed in the preferred embodiment as above, it is not intended to limit the present invention. Anyone with ordinary knowledge in the relevant technical field can make some changes and modifications without departing from the spirit and scope of the present invention. Therefore, the protection scope of the present invention shall be subject to those defined by the attached patent application scope.