CN117812187A - Data processing method and electronic equipment - Google Patents

Data processing method and electronic equipment Download PDF

Info

Publication number
CN117812187A
CN117812187A CN202211214661.5A CN202211214661A CN117812187A CN 117812187 A CN117812187 A CN 117812187A CN 202211214661 A CN202211214661 A CN 202211214661A CN 117812187 A CN117812187 A CN 117812187A
Authority
CN
China
Prior art keywords
content information
scan
information
name
interface
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211214661.5A
Other languages
Chinese (zh)
Inventor
徐超劲
张兰兰
刘石磊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Honor Device Co Ltd
Original Assignee
Honor Device Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Honor Device Co Ltd filed Critical Honor Device Co Ltd
Priority to CN202211214661.5A priority Critical patent/CN117812187A/en
Publication of CN117812187A publication Critical patent/CN117812187A/en
Pending legal-status Critical Current

Links

Abstract

The application provides a data processing method and electronic equipment, and relates to the technical field of terminals. The problem of scan piece seek human-computer interaction inefficiency is solved. The specific scheme is as follows: responding to a first operation of a user, and displaying a first interface; under the condition that a first object is placed in an acquisition area of electronic equipment, responding to the operation of user-indicated scanning, displaying a second interface, wherein the second interface comprises a first scanning image corresponding to the first object; responding to clicking operation of a first control in a second interface by a user, storing a first scanning image, identifying a plurality of pieces of first content information in the first scanning image, and determining first labels corresponding to the plurality of pieces of first content information; in response to an operation indicating to view the stored scan patterns, a third interface is displayed, wherein the third interface includes a plurality of scan result documents including a first scan pattern corresponding to a name determined from second content information in the first scan pattern.

Description

Data processing method and electronic equipment
Technical Field
The present disclosure relates to the field of computer technologies, and in particular, to a data processing method and an electronic device.
Background
During life or work, people use many documents printed on physical media (such as paper), such as contracts, books, periodicals, test papers, and the like. Of course, paper documents are inconvenient to carry and also inconvenient to share over a network.
In the related art, a user can obtain a corresponding scanning piece based on a paper document through an electronic device with a camera, and the corresponding scanning time is used as a name for storage. Therefore, the scanning piece can replace paper documents to be shared in the network, and is convenient for users to carry. However, as the number of scanned parts increases, it is very inconvenient to find the scanned parts that the user wants.
Disclosure of Invention
The embodiment of the application provides a data processing method and electronic equipment, which are used for solving the problem of low man-machine interaction efficiency of searching scanning pieces.
In order to achieve the above purpose, the embodiments of the present application adopt the following technical solutions:
in a first aspect, an embodiment of the present application provides a data processing method, where the electronic device responds to a first operation of a user to display a first interface, and the first interface is a scanning preview interface; under the condition that a first object is placed in an acquisition area of the electronic equipment, the electronic equipment responds to an operation of user indication scanning, and a second interface is displayed, wherein the second interface comprises a first scanning image corresponding to the first object; responding to clicking operation of a first control in the second interface by a user, storing the first scanning image by the electronic equipment, identifying a plurality of pieces of first content information in the first scanning image, and determining a first label corresponding to each piece of first content information in the plurality of pieces of first content information, wherein the first label comprises any one of a header label, a page footer label, a text label, a title label, a graphic label and a single-row text label; after storing the first scan, the electronic device displays a third interface in response to an operation of indicating to view the stored scan, wherein the third interface comprises a plurality of scan result documents, the plurality of scan result documents comprise the first scan, the first scan corresponds to a first name, and the first name is determined according to second content information in the first scan; wherein the second content information is content information satisfying the following condition among the plurality of pieces of first content information: the first label corresponding to the second content information is not a header label, a footer label and a graphic label, the second content information is not watermark information, file type and security level information, and the second content information does not contain preset first keywords.
In the above-described embodiment, the first name configured for the first scan by the electronic apparatus is derived from the content information valid in the first scan, that is, the second content information. In this embodiment, the first name can accurately characterize the content of the first scan while the first name is not affected by extraneous information such as headers, footers, graphics, watermarks, file types, and security levels. Therefore, after the first scanned image is stored, a user can quickly find the first scanned image through the name without clicking all the scanned images, and the man-machine interaction efficiency of searching the scanned images by the user is improved.
In some embodiments, before determining the first tag corresponding to each piece of the first content information in the plurality of pieces of first content information, the method further includes: the electronic equipment determines that the first scanning image is matched with a preset first layout format; the first layout format comprises first position information and second position information, wherein the first position information indicates a first display area of a header in the first scanning diagram, and the second position information indicates a second display area of a footer in the first scanning diagram; the determining the first tag corresponding to each piece of the first content information in the plurality of pieces of first content information includes: when the first content information is located in the first display area, the corresponding first label is the header label; and when the first content information is positioned in the second display area, the corresponding first label is the footer label.
In the embodiment, the electronic device can quickly identify the layout labels corresponding to each content information by using the layout format, so that the identification efficiency and accuracy are improved.
In some embodiments, after displaying the second interface, the method further comprises: when the first scanned image is matched with the first layout format, the electronic equipment displays first prompt information and second prompt information on the second interface, wherein the first prompt information is used for prompting the position of the first display area in the first scanned image, and the second prompt information is used for prompting the position of the second display area in the first scanned image; the electronic equipment responds to the operation of a user on the first prompt information, and the display of the first prompt information is canceled; after the first prompt information is canceled from being displayed, the determining a first tag corresponding to each piece of first content information in the plurality of pieces of first content information includes: canceling the header tag of third content information, the third content information being the first content information located in the first display area; or, the electronic equipment responds to the operation of the user on the second prompt information, and the display of the second prompt information is canceled; after the second prompt information is canceled from being displayed, the determining the first tag corresponding to each piece of first content information in the plurality of pieces of first content information includes: and canceling the header tag of fourth content information, wherein the fourth content information is the first content information positioned in the second display area.
In the above embodiment, the user may manually adjust the first identifier of the content information identified by the electronic device, so as to improve accuracy of identifying the first identifier of the content information.
In some embodiments, the first layout format further corresponds to third location information, where the third location information indicates a third display area of the title in the first scan, and determining the first name according to the second content information in the first scan includes: and if the second content information comprises fifth content information positioned in the third display area under the condition that the first scanning image is matched with the first layout format, the electronic equipment determines the fifth content information as the first name, or extracts a plurality of keywords from the fifth content information and combines the keywords into the first name.
In some embodiments, during display of the second interface, the method further comprises: the electronic device determining a fifth display area in response to a sliding operation of a user, the fifth display area being a contact area with the first scan pattern during the sliding operation of the user; determining the first name according to the second content information in the first scan comprises: when the second content information includes sixth content information located in the fifth display area, the electronic device determines the sixth content information as the first name, or extracts a plurality of keywords from the sixth content information and combines the keywords into the first name.
In some embodiments, determining the first name from the second content information in the first scan comprises: when the second content information comprises seventh content information, the electronic equipment determines the seventh content information as the first name, or extracts a plurality of keywords from the seventh content information to be combined into the first name; wherein the seventh content information includes a preset second category keyword.
In some embodiments, determining the first name from the second content information in the first scan comprises: when the second content information comprises eighth content information, the electronic equipment determines the eighth content information as the first name, or extracts a plurality of keywords from the eighth content information to be combined into the first name;
the eighth content information and the ninth content information correspond to a first difference value, the ninth content information and the tenth content information correspond to a second difference value, the first difference value is larger than the second difference value, the ninth content information is content information except for the eighth content information in the second content information, the tenth content information is content information except for the eighth content information and the ninth content information in the second content information, and the difference value can be determined by a preset font evaluation model based on font sizes, font colors and font special effects among different content information.
In some embodiments, the electronic device is preset with a semantic scoring model, and determining the first name according to the second content information in the first scan includes: the electronic equipment determines a first score corresponding to each piece of second content information by using the semantic score model, wherein the first score indicates the importance degree of the second content information; the electronic equipment determines eleventh content information, wherein the eleventh content information is content information with the highest first score corresponding to the second content information; the electronic device determines the eleventh content information as the first name, or extracts a plurality of keywords from the eleventh content information to be combined into the first name.
In some embodiments, the electronic device is preset with a title generation model, the title generation model is used for summarizing text meaning, the first name is determined according to the second content information in the first scan map, and the method includes: the electronic equipment processes the second content information by using the title generation model and determines a first meaning corresponding to the second content information; the electronic device determines the first meaning as the first name.
In a second aspect, an electronic device provided in an embodiment of the present application includes one or more processors and a memory; the memory is coupled to the processor, the memory for storing computer program code comprising computer instructions that, when executed by the one or more processors, operate to: responding to a first operation of a user, and displaying a first interface, wherein the first interface is a scanning preview interface; under the condition that a first object is placed in an acquisition area of the electronic equipment, responding to the operation of user-indicated scanning, displaying a second interface, wherein the second interface comprises a first scanning image corresponding to the first object; responding to clicking operation of a first control in the second interface by a user, storing the first scanning image, identifying a plurality of pieces of first content information in the first scanning image, and determining a first label corresponding to each piece of first content information in the plurality of pieces of first content information, wherein the first label comprises any one of a header label, a page footer label, a text label, a title label, a graphic label and a single-row text label; after storing the first scan, in response to an operation indicating to view the stored scan, displaying a third interface, wherein the third interface includes a plurality of scan result documents including the first scan, the first scan corresponding to a first name, the first name being a name determined from second content information in the first scan; wherein the second content information is content information satisfying the following condition among the plurality of pieces of first content information: the first label corresponding to the second content information is not a header label, a footer label and a graphic label, the second content information is not watermark information, file type and security level information, and the second content information does not contain preset first keywords.
In some embodiments, before determining a first tag corresponding to each piece of the first content information in the plurality of pieces of first content information, the one or more processors are configured to: determining that the first scanned image is matched with a preset first layout format; the first layout format comprises first position information and second position information, wherein the first position information indicates a first display area of a header in the first scanning diagram, and the second position information indicates a second display area of a footer in the first scanning diagram; when the first content information is located in the first display area, the corresponding first label is the header label; and when the first content information is positioned in the second display area, the corresponding first label is the footer label.
In some embodiments, after displaying the second interface, the one or more processors are configured to: displaying first prompt information and second prompt information on the second interface when the first scan pattern is matched with the first layout format, wherein the first prompt information is used for prompting the position of the first display area in the first scan pattern, and the second prompt information is used for prompting the position of the second display area in the first scan pattern; responding to the operation of a user on the first prompt information, and canceling to display the first prompt information; after the first prompt information is canceled from being displayed, the determining a first tag corresponding to each piece of first content information in the plurality of pieces of first content information includes: canceling the header tag of third content information, the third content information being the first content information located in the first display area; or, in response to the operation of the user on the second prompt information, cancelling the display of the second prompt information; after the second prompt information is canceled from being displayed, the determining the first tag corresponding to each piece of first content information in the plurality of pieces of first content information includes: and canceling the header tag of fourth content information, wherein the fourth content information is the first content information positioned in the second display area.
In some embodiments, the first layout format further corresponds to third location information indicating a third display area of the title in the first scan, and the one or more processors are configured to: and if the second content information comprises fifth content information positioned in the third display area under the condition that the first scanning pattern is matched with the first layout format, determining the fifth content information as the first name, or extracting a plurality of keywords from the fifth content information to be combined into the first name.
In some embodiments, during display of the second interface, the one or more processors are configured to: determining a fifth display area in response to a user's sliding operation, the fifth display area being a contact area with the first scan pattern during the user's sliding operation; when the second content information includes sixth content information in the fifth display area, the sixth content information is determined to be the first name, or a plurality of keywords are extracted from the sixth content information and combined to be the first name.
In some embodiments, determining the first name from the second content information in the first scan comprises: when the second content information comprises seventh content information, determining the seventh content information as the first name, or extracting a plurality of keywords from the seventh content information to be combined into the first name; wherein the seventh content information includes a preset second category keyword.
In some embodiments, the one or more processors are configured to: when eighth content information is included in the second content information, determining the eighth content information as the first name, or extracting a plurality of keywords from the eighth content information to be combined into the first name; a first difference value is corresponding between the eighth content information and the ninth content information, a second difference value is corresponding between the ninth content information and the tenth content information, the first difference value is larger than the second difference value, the ninth content information is content information except for the eighth content information in the second content information, the tenth content information is content information except for the eighth content information and the ninth content information in the second content information, and the difference value can be determined by a preset font evaluation model based on font sizes, font colors and font special effects among different content information.
In some embodiments, the electronic device is pre-configured with a semantic scoring model, and the one or more processors are configured to: determining a first score corresponding to each piece of second content information by using the semantic score model, wherein the first score indicates the importance degree of the second content information; determining eleventh content information, wherein the eleventh content information is content information with the highest corresponding first score in the second content information; and determining the eleventh content information as the first name, or extracting a plurality of keywords from the eleventh content information to be combined into the first name.
In some embodiments, the electronic device is preset with a title generation model for generalizing the meaning, the one or more processors for: processing the second content information by using the title generation model, and determining a first meaning corresponding to the second content information; the first meaning is determined as the first name.
In a third aspect, embodiments of the present application provide a computer storage medium including computer instructions that, when executed on an electronic device, cause the electronic device to perform the method of the first aspect and possible embodiments thereof.
In a fourth aspect, the present application provides a computer program product for, when run on an electronic device as described above, causing the electronic device to perform the method of the first aspect and possible embodiments thereof as described above.
It will be appreciated that the electronic device, the computer storage medium and the computer program product provided in the above aspects are all applicable to the corresponding methods provided above, and therefore, the advantages achieved by the electronic device, the computer storage medium and the computer program product may refer to the advantages in the corresponding methods provided above, and are not repeated herein.
Drawings
Fig. 1 is one of schematic diagrams of a display interface of a scan scene provided in an embodiment of the present application;
FIG. 2 is a schematic diagram of a display interface of a scan scene in the related art;
FIG. 3 is a second schematic diagram of a display interface of a scan scene according to an embodiment of the present disclosure;
FIG. 4 is a third schematic diagram of a display interface of a scan scene according to an embodiment of the present disclosure;
fig. 5 is a diagram illustrating a hardware structure of an electronic device according to an embodiment of the present application;
fig. 6 is a diagram illustrating a software structure of an electronic device according to an embodiment of the present application;
FIG. 7 is a flowchart of a data processing method according to an embodiment of the present application;
FIG. 8 is a schematic diagram of a display interface of a scan scene according to an embodiment of the present disclosure;
FIG. 9 is an exemplary diagram of extracting content information provided in an embodiment of the present application;
FIG. 10 is a fifth schematic diagram of a display interface of a scan scene according to an embodiment of the present disclosure;
FIG. 11 is an exemplary diagram of removing extraneous information provided in an embodiment of the present application;
FIG. 12 is a diagram illustrating a display interface of a scan scene according to an embodiment of the present disclosure;
FIG. 13 is a flowchart of generating a storage name according to an embodiment of the present application;
FIG. 14 is a diagram seventh of a display example of a scan scene provided in an embodiment of the present application;
fig. 15 is an exemplary diagram of a chip system according to an embodiment of the present application.
Detailed Description
The terms "first" and "second" are used below for descriptive purposes only and are not to be construed as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defining "a first" or "a second" may explicitly or implicitly include one or more such feature. In the description of the present embodiment, unless otherwise specified, the meaning of "plurality" is two or more.
The implementation of the present embodiment will be described in detail below with reference to the accompanying drawings.
In the living or working process, people sign contracts, issue journals, issue papers, read books, write test and the like, and usually, the contracts, journals, papers, books and test papers are mostly printed on physical media (such as paper) as original paper, and paper documents are taken as examples in the follow-up. Of course, the user can also utilize the electronic equipment to manufacture the scanning piece corresponding to the original, the scanning piece can be image data and can be stored in the electronic equipment, so that the user can conveniently check at any time, and the user can conveniently upload or share in a network. Of course, the objects that can be scanned may include color pages in a promotional sheet, business cards, documents displayed on an electronic screen (e.g., PPT, PDF, etc.), cards, blackboards, whiteboards, etc., in addition to the paper documents described above.
In some embodiments, a user may use a specialized scanner to make a scan of a corresponding paper document.
In other embodiments, the user may also use an electronic device (e.g., a mobile phone) with a camera to make a scan corresponding to the paper document.
For example, after the electronic device with a camera enables the camera application, a file scanning service provided by the camera application may be invoked to collect image data including an original, and then edge clipping, rectangular correction processing, and other processes are performed on the image data to obtain a corresponding scanned component.
Also, for example, after the electronic device with a camera installs a scanning application, the camera may be called by the scanning application to collect image data including an original, and then edge trimming and rectangular correction processing may be performed on the image data to obtain a corresponding scanned item. The scanning application program may be a third party program that is downloaded in response to a user instruction, or may be a system native application provided in the electronic device, which is not limited in the embodiment of the present application.
Taking fig. 1 as an example, a mobile phone with a scanning application program is installed, and a camera is integrated. After the handset is unlocked, as shown in fig. 1, a main interface 101 is displayed. The main interface 101 includes an application icon, such as icon 102, for scanning an application program. During the period of displaying the main interface 101, when the mobile phone detects that the user selects the icon 102, for example, when the user clicks the display position of the icon 102 in the main interface 101, the mobile phone may start the scanning application program, and display an application interface corresponding to the scanning application program, for example, the scanning main interface 103. The scanning main interface 103 includes an enabling portal for a plurality of functions provided by the scanning application.
Illustratively, the scanning main interface 103 includes a plurality of display controls, each corresponding to a function. For example, the control card 104, the control card 105, the control card 106 in the main interface 103 are scanned. The control card 104 is an entry for enabling the document scanning function, the control card 105 is an entry for enabling the document scanning function, and the control card 106 is an entry for indicating scan data obtained by scanning.
Further, each control card displays information related to the corresponding function, for example, the control card 104 displays a name and an abbreviation of the document scanning function, the control card 105 displays a name and an abbreviation of the document scanning function, and the control card 106 displays the amount of stored scanning data.
When the handset detects operation of the user with respect to the control card 104, the handset may display a scan preview interface, such as interface 107. The interface 107 is a functional interface for scanning paper documents. Wherein the interface 107 includes a scan preview window 108 and a scan confirm control 109. The scan preview window 108 is used for displaying the image frames acquired by the camera, and the scan confirmation control 109 is used for triggering and storing the image frames displayed by the scan preview window 108, for example, referred to as image frame 1, and making a corresponding scan piece (also referred to as scan data) based on the image frame 1.
That is, when the mobile phone switches the display interface 107, the camera is instructed to start, and the camera transmits the acquired image frames back to the scanning application program and displays the image frames on the scanning preview window 108. During this time, the user may place a paper document (e.g., a contract document) that needs to be scanned within the field of view of the camera so that the scan preview window 108 may display the image frames that contain the contract document. In addition, during the display interface 107, the user can also adjust the display position of the contract file in the image frame by adjusting the angle and position of the camera, and the like. In addition, the mobile phone can determine the image frame currently displayed in the scan preview window 108 as the image frame 1 in response to the user's operation of the scan confirm control 109.
In other embodiments, in addition to creating a scan piece based on images captured in real time, a scan piece may be created from stored image data. For example from pictures in a gallery. Of course, the pictures in the gallery may be pictures shot by the mobile phone, or may be pictures downloaded by the mobile phone from the internet, or may be pictures shared by other devices. In summary, the source of the image frame 1 used to make the scan piece is not particularly limited in the embodiments of the present application.
Illustratively, the mobile phone may also respond to the operation of the user on the album import control in the interface 107, and take the picture selected by the user in the mobile phone gallery as the image frame 1. Then, the mobile phone can process the image frame 1 by means of image clipping, rectangular correction and the like to obtain the scanning piece 1. It can be understood that the above-mentioned image clipping may be to clip the image area 1 exclusive to the original from the image frame 1 according to the edge information of the original in the image frame 1, and the specific implementation process may refer to the related art, which is not described herein. The rectangular correction may be to correct the cut-out image area 1 into a rectangular image.
In some embodiments, the process of processing the image frame 1 may not enable rectangular correction, i.e., the image region 1 cut out of the image frame 1 is used as the scan piece 1. In other embodiments, after the image frame 1 is determined, the image frame 1 may be directly used as the scan piece 1 without any processing.
In addition, after the mobile phone obtains the scan 1, the display interface 201 may be switched. The display interface 201 displays the scanner 1. Controls for instructing to adjust the scan 1, such as rescanning controls, watermarking controls, encryption controls, etc., are included in the presentation interface 201.
Wherein the rescanning control is used for indicating that the scanning piece 1 is not saved and the scanning is performed again. For example, when the handset receives a user operation of the rescan control, the handset may redisplay the interface 107.
The watermark control is used to indicate that a specific mark, such as a letter, a pattern, etc., is added to the scanning member 1. The encryption control is used to trigger the addition of access rights to the scanner 1.
In addition, a manual correction control not shown in fig. 1 may also be included in the presentation interface 201. The manual correction control is used to trigger manual rectangular correction for the scan 1. For example, image frame 1 is displayed, and the user is guided to manually mark the edge information of the original in image frame 1, and then, re-capture and rectangular correction are performed.
As shown in fig. 2, a control 202 is also included in presentation interface 201 that indicates the storage scan piece. When the mobile phone receives the operation of the control 202 from the user, the mobile phone can store the scan piece 1 displayed in the display interface 201.
In the related art, as shown in fig. 2, the mobile phone may display an interface 203 in response to a user's operation of a control 202. The interface 203 includes a thumbnail 204 of scan 1, a stored name of scan 1, and a control 205 indicating to view the stored scan. The storage name of the scanning piece 1 is used for distinguishing the scanning piece 1 from other stored scanning pieces.
The name of the scan is usually automatically generated by the mobile phone, and in the related art, the generated name is usually the scan time of the scan, that is, different scan pieces are distinguished by using different scan times.
For example, the acquisition time of scan 1 shown in interface 203 is 2022, 09, 26, 22 and 05 minutes 48, then the mobile phone may name scan 1 as: 2022_09_26_22_05_48, so that the interface 203 can display "new document: 2022_09_26_22_05_48", prompting the user that the scanning element 1 has been named 2022_09_26_22_05_48.
Thereafter, as shown in FIG. 2, the handset may switch to display a "My documents" interface, such as interface 206, in response to user operation of control 205. The interface 206 is used to present all scanned parts that have been scanned. As shown in fig. 2, the interface 206 includes a list window 207. The list window 207 displays information about each stored scanned item, such as an icon or name of each scanned item. However, the names of the respective scan pieces can only suggest that the scan pieces are different, and cannot show the difference between the contents of the scan pieces.
This requires the user to search for the desired target scan, and to start each scan from point to confirm whether it is the target scan. When there are too many scan pieces displayed in the list window 207 and the creation time of the target scan piece is long, the human-machine interaction efficiency of the user to find the target scan piece is low.
In order to improve the above problems, an embodiment of the present application provides a data processing method, which is applied to an electronic device. In this embodiment of the present application, after obtaining the scan data, the electronic device may name the scan data according to the content presented in the scan data, so that when a user browses the scan data stored in the electronic device, the user may locate the target scan data in a large number of stored scan data according to the name of the target scan data.
Continuing with the example in which the electronic device is a mobile phone, after the above data processing method is enabled, as shown in fig. 3, in a scenario in which the mobile phone scans the scan 1, the mobile phone displays the display interface 201. It will be appreciated that during the display of the presentation interface, the scan pieces in the presentation interface are not yet stored. In this scenario, if a click operation of the user on the control 301 indicating the storage is detected, the mobile phone may store the scan 1 displayed in the presentation interface 201, and may determine that the name of the scan 1 is a labor contract according to the content of the scan 1, and then the mobile phone switches the display interface 302. The interface 302 includes a thumbnail of the scan 1, a name (labor contract) of the scan 1, and a control 303 indicating to view the stored scan. It will be appreciated that the stored scanner includes scanner 1. In this scenario, the handset may switch to display a "My documents" interface, such as interface 304, in response to user operation of control 303. The interface 304 includes a list window 305. The list window 305 displays information about each stored scanned item, such as an icon or name of each scanned item. In addition, each scan displayed in the list window 305 has a selection control displayed correspondingly, for example, scan 1 (named labor contract) has a selection control 306 displayed in the list window 305, and the mobile phone can determine that the user has selected scan 1 in response to the user selecting the selection control 306.
In this scenario, if the user wants to perform subsequent processing on the scan 1, such as sharing, deleting, merging, printing, etc., the scan 1 can be quickly found from the list window 305 by the name (i.e., labor contract) of the scan 1. The manner of searching the scan element 1 may be manual searching, or searching may be performed by inputting words appearing in the corresponding stored names through a search field, which is not specifically limited in the embodiment of the present application.
Thus, after the handset detects that the user clicks the select control 306, it can be determined that the user selected scan 1. After the scanning piece 1 is selected, the mobile phone can perform corresponding processing on the scanning piece 1 according to the operation of the user.
Illustratively, after determining that scan 1 is selected, the handset may send scan 1 to the other device in response to a user operation of the sharing control in interface 304. Illustratively, after determining that scan 1 is selected, the handset may delete scan 1 in response to a user operation of a delete control in interface 304. Illustratively, after determining that scan 1 and scan 2 (named six-level lower school unit test) are selected, the handset may merge scan 1 and scan 2 into one data packet in response to user operation of the merge control in interface 304. Illustratively, after determining that scan 1 is selected, the handset may send scan 1 to the printing device in response to a user operation of the print control in interface 304.
Additionally, in some possible embodiments, the handset may skip displaying the presentation interface 201, so that the interface 302 may be displayed directly after the handset receives an operation from the user to instruct scanning (e.g., the user clicks the scan confirm control 109). Of course, the mobile phone may display the presentation interface 201, and after receiving the operation of the user to instruct saving, the mobile phone may switch back to the scan preview interface, the application main interface of the scan application, or the like, instead of displaying the interface 302.
Obviously, when the mobile phone scans originals with different contents, the names of the obtained scanned parts are different. For example, as shown in fig. 4, in a case where the mobile phone displays the interface 107 again in response to the user operation, after the user places the original of the additional terms of the labor contract in the field of view of the camera, the image frame in which the original of the additional terms of the labor contract is displayed is included in the interface 107. The handset then displays a presentation interface 401 as shown in fig. 4 in response to user operation of the scan confirm control 109. Wherein the presentation interface 401 is similar to the presentation interface 201, with the difference that the scan displayed in the presentation interface 401 is scan 3 (i.e., additional terms of labor contracts). In response to a user's operation of the confirm store control 402, the handset may determine the storage name of the scan 3 from the content in the scan 3, e.g., determine that the storage name of the scan 3 is an additional term of the labor contract. Thereafter, the handset may display interface 403. Among these are a thumbnail of the scan 3, a storage name (labor contract additional term) of the scan 3, and a control 404 indicating to view the stored scan. In the case that the mobile phone detects that the user indicates to select the control 404, the mobile phone can switch and display an interface such as an interface 405 of my document, and compared with the interface 304 in fig. 3, the interface 405 has an icon and a name of the scanning part 3 added, and meanwhile, the corresponding storage names of the scanning part 3 and the scanning part 1 are different due to different contents.
In addition, the scenario where the mobile phone displays the interface 107 again in response to the user operation may be that the mobile phone detects the operation of switching back to the scan main interface 103 with the instruction, and then detects the operation of clicking the control card 104 by the user.
In some embodiments, where the scan main interface 103 is displayed, the handset may also display a "my documents" interface containing stored scan pieces, e.g., interface 304, interface 405, in response to user operation of the control card 106.
The method is that the scanning application program is adopted to create the scanning piece and name the scanning piece, and the same method can be adopted to name the obtained scanning piece when the camera application in the mobile phone starts the scanning file function, so that a user can conveniently and quickly find the target scanning piece in the gallery.
In some embodiments, the mobile phone names the scanned document according to the content printed in the certificate, and the specific process is similar to the process of scanning and naming the paper document, in the subsequent embodiments, the scanning and naming are mainly taken as examples for the paper document, but it is understood that the method is also applicable to color pages, business cards, documents displayed on an electronic screen (such as PPT, PDF, etc.), cards, blackboards, whiteboards, and other types of scanned objects in a publicity sheet.
In addition, after the scanned part is scanned, the determined storage name is a name related to the scanned part, and when the user instructs the mobile phone to generate an output file according to the scanned part, for example, files in other formats (such as PDF, PPT or word) are made based on the scanned part, the name of the output file can be determined by using the storage name corresponding to the scanned part. For example, when an output file is created from a single scan, the stored name of the scan is directly used as the name of the output file. For example, when an output file in PDF format is created from the scanner 1, "labor contract" may be used as the name of the output file in PDF format.
Also, for example, when an output file is created from a plurality of scan pieces, the stored names of the plurality of scan pieces may be combined as the names of the output file. For example, when an output file in PDF format is created from the scanner 1 and the scanner 3, "labor contract and labor contract additional terms" may be used as the names of the output files in PDF format.
In addition, when the user instructs to combine a plurality of scan pieces into a combined scan piece, if the stored names of the combined scan pieces are the same, the stored name of the scan piece is directly used as the name of the combined scan piece. If the combined scanned item storage names are different, the different storage names are combined as the combined scanned item names. For example, when the scanner 1 and the scanner 3 are combined, the "labor contract and the additional term of the labor contract" may be used as the name of the combined scanner.
Of course, the electronic device may be other devices besides a mobile phone. The electronic device may also be a tablet, desktop, laptop, handheld, notebook, personal computer (personal computer, PC), netbook, cellular telephone, personal digital assistant (personal digital assistant, PDA), wearable device (e.g., smart watch), car computer, game console, and augmented reality (augmented reality, AR) \virtual reality (VR) device, for example. The embodiment of the application does not particularly limit the specific form of the electronic device.
Taking an electronic device as a mobile phone example, fig. 5 shows a schematic structural diagram of the mobile phone.
The handset may include a processor 110, an external memory interface 120, an internal memory 121, a universal serial bus (universal serial bus, USB) interface 130, an antenna 1, an antenna 2, a mobile communication module 150, a wireless communication module 160, an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, a sensor module 180, and the like.
It will be appreciated that the structure illustrated in the embodiments of the present invention is not limited to a specific configuration of the mobile phone. In other embodiments of the present application, the handset may include more or fewer components than shown, or certain components may be combined, or certain components may be split, or different arrangements of components may be provided. The illustrated components may be implemented in hardware, software, or a combination of software and hardware.
The processor 110 may include one or more processing units, such as: the processor 110 may include an application processor (application processor, AP), a modem processor, a graphics processor (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), a controller, a memory, a video codec, a digital signal processor (digital signal processor, DSP), a baseband processor, and/or a neural network processor (neural-network processing unit, NPU), etc. Wherein the different processing units may be separate devices or may be integrated in one or more processors.
A memory may also be provided in the processor 110 for storing instructions and data. In some embodiments, the memory in the processor 110 is a cache memory. The memory may hold instructions or data that the processor 110 has just used or recycled. If the processor 110 needs to reuse the instruction or data, it can be called directly from the memory. Repeated accesses are avoided and the latency of the processor 110 is reduced, thereby improving the efficiency of the system.
The wireless communication function of the mobile phone can be realized by the antenna 1, the antenna 2, the mobile communication module 150, the wireless communication module 160, a modem processor, a baseband processor and the like.
The antennas 1 and 2 are used for transmitting and receiving electromagnetic wave signals. Each antenna in the handset may be used to cover a single or multiple communication bands. Different antennas may also be multiplexed to improve the utilization of the antennas. For example: the antenna 1 may be multiplexed into a diversity antenna of a wireless local area network. In other embodiments, the antenna may be used in conjunction with a tuning switch.
The mobile communication module 150 may provide a solution for wireless communication including 2G/3G/4G/5G, etc. applied to a cell phone. The mobile communication module 150 may include at least one filter, switch, power amplifier, low noise amplifier (low noise amplifier, LNA), etc. The mobile communication module 150 may receive electromagnetic waves from the antenna 1, perform processes such as filtering, amplifying, and the like on the received electromagnetic waves, and transmit the processed electromagnetic waves to the modem processor for demodulation. The mobile communication module 150 can amplify the signal modulated by the modem processor, and convert the signal into electromagnetic waves through the antenna 1 to radiate. In some embodiments, at least some of the functional modules of the mobile communication module 150 may be disposed in the processor 110. In some embodiments, at least some of the functional modules of the mobile communication module 150 may be provided in the same device as at least some of the modules of the processor 110.
The wireless communication module 160 may provide solutions for wireless communication including wireless local area network (wireless local area networks, WLAN) (e.g., wireless fidelity (wireless fidelity, wi-Fi) network), bluetooth (BT), global navigation satellite system (global navigation satellite system, GNSS), frequency modulation (frequency modulation, FM), near field wireless communication technology (near field communication, NFC), infrared technology (IR), etc. applied to a cell phone. The wireless communication module 160 may be one or more devices that integrate at least one communication processing module. The wireless communication module 160 receives electromagnetic waves via the antenna 2, modulates the electromagnetic wave signals, filters the electromagnetic wave signals, and transmits the processed signals to the processor 110. The wireless communication module 160 may also receive a signal to be transmitted from the processor 110, frequency modulate it, amplify it, and convert it to electromagnetic waves for radiation via the antenna 2.
In some embodiments, the antenna 1 and the mobile communication module 150 of the handset are coupled, and the antenna 2 and the wireless communication module 160 are coupled, so that the handset can communicate with a network and other devices through wireless communication technology. The wireless communication techniques may include the Global System for Mobile communications (global system for mobile communications, GSM), general packet radio service (general packet radio service, GPRS), code division multiple access (code division multiple access, CDMA), wideband code division multiple access (wideband code division multiple access, WCDMA), time division code division multiple access (time-division code division multiple access, TD-SCDMA), long term evolution (long term evolution, LTE), BT, GNSS, WLAN, NFC, FM, and/or IR techniques, among others. The GNSS may include a global satellite positioning system (global positioning system, GPS), a global navigation satellite system (global navigation satellite system, GLONASS), a beidou satellite navigation system (beidou navigation satellite system, BDS), a quasi zenith satellite system (quasi-zenith satellite system, QZSS) and/or a satellite based augmentation system (satellite based augmentation systems, SBAS).
The cell phone implements display functions through the GPU, the display 194, and the application processor, etc. The GPU is a microprocessor for image processing, and is connected to the display 194 and the application processor. The GPU is used to perform mathematical and geometric calculations for graphics rendering. Processor 110 may include one or more GPUs that execute program instructions to generate or change display information.
The display screen 194 is used to display images, videos, and the like. The display 194 includes a display panel. The display panel may employ a liquid crystal display (liquid crystal display, LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode (AMOLED) or an active-matrix organic light-emitting diode (matrix organic light emitting diode), a flexible light-emitting diode (flex), a mini, a Micro led, a Micro-OLED, a quantum dot light-emitting diode (quantum dot light emitting diodes, QLED), or the like. In some embodiments, the handset may include 1 or N display screens 194, N being a positive integer greater than 1.
The cell phone may implement photographing functions through an ISP, a camera 193, a video codec, a GPU, a display 194, an application processor, and the like.
The ISP is used to process data fed back by the camera 193. For example, when photographing, the shutter is opened, light is transmitted to the camera photosensitive element through the lens, the optical signal is converted into an electric signal, and the camera photosensitive element transmits the electric signal to the ISP for processing and is converted into an image visible to naked eyes. ISP can also optimize the noise, brightness and skin color of the image. The ISP can also optimize parameters such as exposure, color temperature and the like of a shooting scene. In some embodiments, the ISP may be provided in the camera 193.
The camera 193 is used to capture still images or video. The object generates an optical image through the lens and projects the optical image onto the photosensitive element. The photosensitive element may be a charge coupled device (charge coupled device, CCD) or a Complementary Metal Oxide Semiconductor (CMOS) phototransistor. The photosensitive element converts the optical signal into an electrical signal, which is then transferred to the ISP to be converted into a digital image signal. The ISP outputs the digital image signal to the DSP for processing. The DSP converts the digital image signal into an image signal in a standard RGB, YUV, or the like format. In some embodiments, the handset may include 1 or N cameras 193, N being a positive integer greater than 1.
The digital signal processor is used for processing digital signals, and can process other digital signals besides digital image signals. For example, when the mobile phone selects a frequency point, the digital signal processor is used to perform fourier transform on the frequency point energy, etc.
Video codecs are used to compress or decompress digital video. The handset may support one or more video codecs. In this way, the mobile phone can play or record videos with various coding formats, for example: dynamic picture experts group (moving picture experts group, MPEG) 1, MPEG2, MPEG3, MPEG4, etc.
The external memory interface 120 may be used to connect to an external memory card, such as a Micro SD card, to extend the memory capabilities of the handset. The external memory card communicates with the processor 110 through an external memory interface 120 to implement data storage functions. For example, files such as music, video, etc. are stored in an external memory card.
The internal memory 121 may be used to store computer executable program code including instructions. The processor 110 executes various functional applications of the cellular phone and data processing by executing instructions stored in the internal memory 121. The internal memory 121 may include a storage program area and a storage data area. The storage program area may store an application program (such as a sound playing function, an image playing function, etc.) required for at least one function of the operating system, etc. The storage data area may store data created during use of the handset (e.g., audio data, phonebook, etc.), etc. In addition, the internal memory 121 may include a high-speed random access memory, and may further include a nonvolatile memory such as at least one magnetic disk storage device, a flash memory device, a universal flash memory (universal flash storage, UFS), and the like.
The handset may implement audio functions through an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, an application processor, and the like. Such as music playing, recording, etc.
Wherein the audio module 170 is used to convert digital audio information into an analog audio signal output and also to convert an analog audio input into a digital audio signal. The audio module 170 may also be used to encode and decode audio signals. In some embodiments, the audio module 170 may be disposed in the processor 110, or a portion of the functional modules of the audio module 170 may be disposed in the processor 110.
Microphone 170C, also referred to as a "microphone" or "microphone", is used to convert sound signals into electrical signals. When making a call or transmitting voice information, the user can sound near the microphone 170C through the mouth, inputting a sound signal to the microphone 170C. The handset may be provided with at least one microphone 170C. In other embodiments, the handset may be provided with two microphones 170C, which may also perform noise reduction in addition to collecting sound signals. In other embodiments, the handset may also be provided with three, four or more microphones 170C to enable collection of sound signals, noise reduction, identification of sound sources, directional recording, etc.
Of course, the mobile phone may further include a charging management module, a power management module, a battery, a key, an indicator, 1 or more SIM card interfaces, and the embodiment of the present application does not limit the present application.
In addition, the mobile phone also comprises a software system. The software system of the mobile phone can adopt a layered architecture, an event driven architecture, a microkernel architecture, a microservice architecture or a cloud architecture.
The layered architecture described above divides the software into several layers, each with distinct roles and branches. The layers communicate with each other through a software interface. In some embodiments, as shown in fig. 6, the software system of the mobile phone is divided into multiple layers, namely, an application layer, an application framework layer, an algorithm engine layer, an Zhuoyun row (Android run) and system library, and a kernel layer from top to bottom.
The application layer may include a series of application packages.
Illustratively, the application layer may include: gallery, camera, scan, memo, etc. APP, although other applications not shown in the figures may be included.
The application framework layer provides an application programming interface (application programming interface, API) and programming framework for application programs of the application layer. The application framework layer includes a number of predefined functions.
Illustratively, the application framework layer may include: print manager, window manager, resource manager, notification manager, activity manager, etc.
Wherein the print manager is a print framework provided by the mobile phone, and the PrintJob print class provided by the print manager can be used for creating a print job, that is, converting content to be printed into a printable document. In addition, the print manager may send a document to be printed (e.g., a print job) to the connected printing apparatus to instruct the printing apparatus to complete the print job. In addition, a print adapter is included in the application framework layer that mates with the print manager for selecting the printer and print options.
The window manager is used for managing window programs. The window manager can acquire the size of the display screen, judge whether a status bar exists, lock the screen, intercept the screen and the like.
The resource manager provides various resources, such as localization strings, icons, pictures, layout files, video files, and the like, to the application program.
The notification manager enables the application to display notification information in a status bar, can be used for conveying notification type messages, and can automatically disappear after a short stay without user interaction. Such as notification manager is used to inform that the download is complete, message alerts, etc. The notification manager may also be a notification in the form of a chart or scroll bar text that appears on the system top status bar, such as a notification of a background running application, or a notification that appears on the screen in the form of a dialog window. For example, a text message is presented in a status bar, a prompt tone is emitted, vibration is generated, and an indicator light blinks.
Also for example, the system library may include a plurality of functional modules. For example: layer integrators (surfeflingers), media Libraries (Media Libraries), three-dimensional graphics processing Libraries (e.g., openGL ES), 2D graphics engines (e.g., SGL), etc. SurfaceFlinger is used to manage the display subsystem and provides a fusion of 2D and 3D layers for multiple applications. Media libraries support a variety of commonly used audio, video format playback and recording, still image files, and the like. The media library may support a variety of audio and video encoding formats, such as MPEG4, h.264, MP3, AAC, AMR, JPG, PNG, etc. The three-dimensional graphic processing library is used for realizing three-dimensional graphic drawing, image rendering, synthesis, layer processing and the like. The 2D graphics engine is a drawing engine for 2D drawing.
The algorithm engine layer comprises a document content identification and layout analysis module, a content analysis module and a file name extraction/generation module.
The document content recognition and layout analysis module recognizes and extracts the content of the scanned piece through optical character recognition (Optical Character Recognition, OCR). In addition, layout analysis enables identification of component categories: including identifying the header, footer, header, paragraph, line, character, graphic, etc. of the document.
And the content analysis module is used for performing step analysis on the main body content according to the identification result output by the document content identification and layout analysis module, and extracting information which is not greatly related to the actual content of the scanned main body, such as watermark, number, file type, confidentiality level and the like. Not limited to text classification techniques (deep learning networks) as identified by the above text categories, or using keyword judgment techniques. This step marks the content that has not been identified to have a great relevance to the subject content.
And the file name extraction/generation module is used for analyzing the content of the scanned piece, extracting/generating proper information and taking the proper information as a default name stored in the scanned piece.
Android run time includes a core library and virtual machines. Android run time is responsible for scheduling and management of the Android system.
The core library consists of two parts: one part is a function which needs to be called by java language, and the other part is a core library of android.
The application layer and the application framework layer run in a virtual machine. The virtual machine executes java files of the application program layer and the application program framework layer as binary files. The virtual machine is used for executing the functions of object life cycle management, stack management, thread management, security and exception management, garbage collection and the like.
The kernel layer is a layer between hardware and software. The kernel layer at least comprises a camera driver, an audio driver, a sensor driver and the like, and the embodiment of the application does not limit the application.
According to the method provided by the embodiment of the application layer, the camera (namely, the optical lens) can be called to collect the image frame containing the original through the document scanning function applied by the camera in the application layer or the scanning application program of the application layer, and the scanned part is obtained through processing. Then, the scanned document is subjected to OCR (optical character recognition) and version analysis by a document content recognition and layout analysis engine, text information and component types are extracted, and the content analysis engine further analyzes information which is not greatly related to the actual content of the scanned document body, such as watermark, number, file type, confidentiality level and the like. And finally, sending the identified data to a file name extraction/generation module to generate a file name.
The hardware structure and the software system construct an operation environment for realizing the technical scheme provided by the embodiment of the application. The following describes a technical solution in an embodiment of the present application by taking an electronic device as an example of a mobile phone and referring to the accompanying drawings.
In some embodiments, as shown in fig. 7, the above method may include:
S101, the mobile phone acquires the scanning piece a.
In some embodiments, as shown in fig. 1, the mobile phone may obtain a scanned item a, that is, a scanned item corresponding to a paper labor contract by enabling a file scanning function in a scanning application. In other embodiments, as shown in FIG. 8, the handset may enable a camera application in response to user operation. In the capture preview interface 801 provided by the camera application, the user can select a document scanning function provided by the camera application in a sliding manner. In this scenario, the handset may display a preview interface 802 for scanning in response to a user selecting the document scanning function. Then, the mobile phone can respond to the operation of the user for indicating acquisition to obtain a scanning piece a, and display a display interface 803, wherein the display interface 803 comprises the scanning piece a.
S102, the mobile phone identifies content information a appearing in the scanning piece a.
The content information a may be all content except a background layer in the scan a, for example, the content information a may include characters, marks, graphics, and the like appearing in the scan a.
As an implementation manner, the mobile phone can use OCR technology to identify the scanned item a and extract the corresponding content information a. For example, as shown in fig. 9, after the scan a is processed by the OCR module, the output content information a includes: "9 months of 2022", "XX department", horizontal line, "article number: 12345abc ", secret level," secret "," notification about xxxx ", a piece of text, stamp, and" internal file ".
In some embodiments, the mobile phone can also acquire the position, arrangement, occupation display area, format and other information of each piece of content information a in the scanning piece a while identifying the content information a through the OCR technology. Then, according to the position, arrangement, format and other information of the content information a in the scanning piece a, determining the layout label corresponding to the content information a. Wherein, the layout label includes: header, footer, title, paragraph, single line, character, image, etc.
In the first embodiment, the mobile phone may determine the tag corresponding to the content information a in combination with the file format selected in advance. That is, a plurality of file formats may be stored in the mobile phone in advance, and each file format corresponds to one file format. For example, journal formats, paper formats, contract formats, bidding document formats, report formats, briefcase formats, book cover formats, promotional page formats, book content formats, and the like may be included.
In addition, the step of selecting the file format may be performed before the mobile phone scans the scan a.
For example, as shown in fig. 10, during the display of the capture preview interface 1001, the user may select the document scanning function provided by the camera application by sliding. In this scenario, the handset may display a preview interface 1002 for scanning in response to a user selecting the document scanning function. A plurality of selectable file format identifiers, such as "document tab", "journal paper tab", "book cover tab", "contract agreement tab", etc., are included in the preview interface 1002, and of course, file format tabs not shown in the figure may be provided.
In some embodiments, each file layout may correspond to one layout information, which illustratively includes the positions and sizes of a plurality of layout areas, such as the positions and sizes of a header display area, a footer display area, a text display area, and the like.
As further shown in FIG. 10, after determining that the user selected "document tab" and indicated that the scan was initiated, scanner 1003 may be obtained. The layout area distribution of the document comprises position and size information of a header display area, a title display area and a text display area. When the scanner 1003 is obtained and it is determined that the scanner 1003 corresponds to a document, a header display area (area 1004), a title display area (area 1005), and a text display area (area 1006) may be divided in the scanner 1003 according to the layout area distribution of the document.
In this way, after the content information is extracted from the scanner 1003, the content information in the area 1004 may be given as a header tag, the content information in the area 1005 may be given as a title tag, and the content information in the area 1006 may be given as a text tag, and of course, the content information is not extracted in the area 1006 of the scanner 1003.
In addition to selecting a file format before scanning, a corresponding file format may also be selected after scanning (e.g., during displaying a display interface including the scanned item), so that when scanning is completed and the display interface including the scanned item is displayed, the mobile phone may determine, in response to a user operation, a file format selected by the user for the current scanned item.
Under the condition that the file format is selected, the mobile phone can assign a corresponding layout label to the extracted content information according to whether the area where the extracted content information is located is a header display area, a footer display area, a title display area or a text display area. If the extracted content information does not belong to the above-described area, the layout label thereof may be given according to other features. For example, if the content information is graphics, then a graphic tag is given thereto. For example, if the content information is a line of characters, then a single line of labels is assigned thereto. For another example, if the content information is a paragraph, then a paragraph label is assigned to the content information. Alternatively, if the acquired content information does not belong to any of the above-described areas, the layout label may not be given.
In addition, for content information appearing in the text display area, it is also necessary to discriminate whether text labels can be given, for example, content information in which the graphics and text directions appearing in the text display area are inconsistent with the main paragraph may not be given labels.
In other embodiments, during the display of the scanning element 1003 in the display interface, the mobile phone may also display the edge information of the header display area, the title display area, and the text display area in the scanning element 1003, and display a corresponding identifier on each edge information, for example, display the identifier "header" on the edge of the header display area. Therefore, the user can intuitively see the partition of the mobile phone to the layout of the scanned part, and can judge whether the partition is accurate or not. If the mobile phone receives an operation to cancel the header display area at this time, a header label is not given even if there is content information a in the area 1004. If the mobile phone receives an operation of moving the header display area at this time, the content information a located in the moved header display area may be given a header tag.
In the second implementation manner, if the user does not select the file format, the layout label corresponding to each piece of content information may also be determined in the following several ways:
(1) And determining different layout areas according to the marking characteristics of the indication layout areas.
For example, as shown in fig. 9, the distance between the "horizontal line" of the content information a and the upper edge of the scan piece a is smaller than a preset distance value, and the "horizontal line" is identified as a header line, and a header label of the "horizontal line" may be given. The area between the upper edge of the scanning member a and the header line is determined as a header display area, and the header label is given to all the content information appearing in the header display area. For example, the content information a "9 nd month 2022" and "XX department" is located between the upper edge of the scanner a and the header line, and the content "9 nd month 2022" and "XX department" are identified as the contents of the header and are given a header label.
Also, for example, the mobile phone determines the page number in the scan piece, and may determine the footer display area according to the position where the page number is located, and assign the content information appearing in the footer display area to the footer tag.
(2) And determining the matched file formats from the prestored multiple file formats by using the similar identification model. Thus, a plurality of corresponding layout areas can be determined according to the matched file formats, and accordingly layout labels corresponding to the content information can be determined.
(3) And determining the corresponding layout label according to the characteristics of the content information.
Illustratively, as shown in fig. 9, "article number: 12345abc "," confidential "," notification about xxxx ", and" internal file "are all single line characters that run horizontally, which may be assigned a single line label. The "stamp" includes various complex curves to which graphic labels may be assigned. The character of a paragraph corresponds to the characteristics of first line indentation, two sides alignment, punctuation marks and the like, and can be endowed with paragraph labels. Since the layout and direction of the "secret" are different from those of a single line of text, the corresponding layout label cannot be determined, and in this case, the content information layout label may not be given.
In other embodiments, the same content information may also include multiple layout tags, e.g., "notification about xxxx" may include both a title tag and a single row tag.
The implementation manner of determining the layout label corresponding to the content information according to the position, arrangement, format and other information of the content information in the scanning piece is not limited to determining the layout label corresponding to each content information by adopting the method in the embodiment of the application.
As an embodiment, S102 may be performed by a document content recognition and layout analysis module.
S103, the mobile phone screens out irrelevant information in the content information a according to the layout label of the content information a and preset conditions to obtain content information b.
The content information meeting the preset condition belongs to irrelevant information, and the preset condition can be preset or changed in response to the operation of a user. In some embodiments, the preset conditions may include at least one of:
(1) The content information a contains any item of tag such as a header tag, a footer tag, a graphic tag and the like, and belongs to irrelevant information.
(2) The content information of the content information a, the content semantic non-conforming requirement, also belongs to irrelevant information.
For example, sample information, such as number sequence, file type, security level, may be preconfigured for semantic disagreement requirements. Then, the mobile phone can identify the content semantics corresponding to the content information a by using the semantic identification model. Then, the similarity between the content semantics of the content information a and the content semantics of the sample information is evaluated. And under the condition that the similarity exceeds a preset similarity threshold value 1, determining that the semantics of the content information a are not satisfactory. The semantic recognition model may be a model obtained by training a neural network model, and the specific training process may refer to related technologies and will not be described herein. In addition, the mobile phone can also determine the similarity between different content semantics by using a semantic similarity recognition model. The semantic similarity recognition model can also be obtained by training a neural network model, and the training process can refer to the related technology, and is not described herein.
On the basis of the above example, the sample information may be associated with a file format, and of course, the same sample information may be associated with a plurality of file formats, so that the same sample information may exist between different file formats, or different sample information may exist between different file formats. The number of sample information corresponding to different file formats may be the same or different. The embodiment of the present application is not particularly limited thereto. On the basis, when the mobile phone determines that the scanned piece corresponds to a file format (for example, the file format which is manually selected by a user is determined, and for example, the mobile phone determines a matched file format through a similar recognition model), whether the content semantics of the content information are not in accordance with the requirements can be judged according to sample information corresponding to the file format, so that the workload of semantic comparison is reduced to a certain extent, and the pertinence is improved.
After identifying the content information c with the text label and the paragraph label, extracting the content semantic 1 corresponding to the content information c by using the semantic identification model, extracting the content semantic 2 corresponding to other content information by using the semantic identification model, and determining whether the content information corresponding to the content semantic 2 belongs to irrelevant information by evaluating the similarity between the content semantic 2 and the content semantic 1. For example, the similarity is lower than the preset similarity threshold 2, and it may be determined that the content information corresponding to the content semantic 2 belongs to irrelevant information. The preset similarity threshold 2 is smaller than the preset similarity threshold 1.
(3) The content information a including the keyword 1 also belongs to irrelevant information.
The keyword 1 may be a pre-configured word, and the keyword 1 may be a word frequently appearing in irrelevant information marked by a user, or may be a word directly specified by the user. For example, the class of confidentiality, confidential, watermark related vocabulary, number, document number, date vocabulary (e.g., year, month, day), etc., and the type of the corresponding keyword 1 is not particularly limited.
In other embodiments, keyword 1 may also be associated with a file layout. Of course, the same keyword 1 may be associated with multiple file formats. Thus, different file formats may correspond to different keywords 1, or may correspond to the same keywords 1. Thus, after determining the file format corresponding to the scanned item, the keyword 1 corresponding to the file format can be used to determine irrelevant information from the content information corresponding to the scanned item.
(4) The irrelevant information is identified from the content information a by a text classification technique.
In some embodiments, the text classification technique may be to identify content belonging to the watermark, the number, the file type, the security level, and the like from the content information a as irrelevant information by using a neural network model.
In other possible embodiments, the layout information corresponding to each file format may further include an area for displaying irrelevant information. Thus, after determining the text format corresponding to the scanned item, if the layout information of the file format includes an area for displaying irrelevant information and the scanned item has content information located in the area for displaying irrelevant information, the content information may be used as irrelevant information.
In some embodiments, after determining the extraneous information in content information a, the extraneous information is filtered out to obtain content information b.
For example, as shown in fig. 11, in the content information a corresponding to the scanner a, "2022, 9 th month", "XX department", and a horizontal line have header labels, and are defined as irrelevant information. The "stamp" has a graphic label and is also considered as extraneous information. Article number: 12345abc ", and the content information a such as the confidentiality level has the keyword 1 (e.g., number, confidentiality level), and is also regarded as irrelevant information. In addition, through text classification technology, it is identified that "secret" belongs to the watermark, and "internal file" belongs to the file type, which can be regarded as irrelevant information. Then, after screening out the above irrelevant information, the obtained content information b may include "notification about xxxx" and a piece of text.
While various conditions for determining invalid information are described above, it will be appreciated that in various embodiments, one or more of the above conditions may be selected as the conditions for determining invalid information, and of course, other conditions for determining invalid information, which are not described, may be included, which is not particularly limited to the embodiments of the present application.
In addition, in some possible embodiments, after the mobile phone identifies all irrelevant information, only part of irrelevant information can be screened out, so that the processing efficiency is improved. For example, invalid information arranged above the central axis of the scan piece a is screened out. Of course, the identified invalid information may be completely screened out, which is not particularly limited in the embodiment of the present application.
In other embodiments, if no extraneous information is contained in the content information a, then all the content information a is determined to be content information b.
As an embodiment, the above S103 may be performed by a content analysis module.
S104, the mobile phone determines the storage name corresponding to the scanning piece a according to the content information b.
In some embodiments, the storage name may be determined using the arrangement position of the content information b in the scan a. The stored name is a name associated with the scanner a itself, and is not limited to a name derived from the scanner into a file of another format.
For example, the content information for extracting the storage name is determined according to the height of the content information b in the scanning piece a, wherein the higher the height value corresponding to the content information b near the upper side of the scanning piece a. For example, as shown in fig. 12, in the scan a, "notification about xxxx" is displayed in the area 1201, a text is displayed in the area 1202, and it is obvious that the area 1201 is further on the upper side, that is, the height value corresponding to "notification about xxxx" is highest, and in this scenario, the mobile phone may extract the storage name according to "notification about xxxx".
In other possible examples, the handset may also obtain content information b (e.g., referred to as content information d) with a height value arranged before a specified ranking (e.g., 2), and then determine a relevance score for each piece of content information d to the scanned piece of content via an artificial intelligence model. And selecting content information d with highest score, and extracting a storage name according to the content information d. In the foregoing example, when the content information b of the maximum height value has a plurality of pieces, the content information for extracting the storage name may be determined in the same manner.
For example, the content information for extracting the storage name is determined from the display area to which the content information b belongs in the scanner a. For example, after determining the file format corresponding to the scanned item a (detecting that the user selects the corresponding file format for the scanned item a, or identifying the file format corresponding to the scanned item a), the mobile phone may determine content information for extracting the storage name according to the title display area of the scanned item a in the file format. That is, if a piece of content information b appears in the title display area, the storage name may be determined based on the content information b.
In the above embodiment, if the content information b is not contained in the title display region, the upper edge of the tag display region is set as the start point, the content information b closest to the upper edge is searched upward, and the lower edge of the tag display region is set as the start point, the content information b closest to the lower edge is searched downward. In some examples, the above-mentioned distance may also be understood as the smallest difference in height.
It should be noted that, the manner of extracting the storage name from the content information may include:
(1) Directly take it as the storage name. For example, the content information for extracting the storage name is "notification about xxxx", and "notification about xxxx" may be directly used as the storage name.
(2) And extracting a plurality of keywords 2 from the content information, wherein the keywords 2 are words close to corresponding content semantics in the content information, and then combining the extracted keywords 2 to obtain a storage name. For example, if the content information for extracting the storage name is "notification about xxxx", and then the above-described "xxxx" and "notification" are determined as the keyword 2, then "xxxx notification" may be used as the storage name.
In some embodiments, the content information for extracting the storage name may also be determined based on the character information of the content information b.
For example, a plurality of keywords 3 are preset in the mobile phone, and the preset keywords 3 may be words commonly appearing in the document theme, for example, the keywords 3 include "title", "topic", "conclusion", "summary", and the like. If the scanned item a contains a plurality of pieces of content information b having the keyword 3, the content information b having the keyword 3 may also be scored, and the score may be a correlation score between the content information and the scanned item content. Then, the content information b having the keyword 3 and having the highest score is selected as the content information for extracting the storage name.
Also for example, a font evaluation model may be configured in the mobile phone, and the font evaluation model may evaluate a difference value between different content information according to font sizes, font colors, font special effects, and other font types of the different content information. The mobile phone can recognize a difference value between each piece of content information b and each piece of other content information b using the font evaluation model. Then, the content information (e.g., referred to as content information e) having the largest difference in font type from other content information is searched for using the difference value, and then the storage name of the scanner a is extracted based on the content information e.
For example, the content information 1 and the content information 2 in the content information b correspond to the difference value 1, and the content information 3 and the content information 2 in the content information b correspond to the difference value 2. Wherein the content information 2 is content information other than the content information 1 in the content information b, the content information 3 is content information other than the content information 1 and the content information 2 in the content information b, and if the difference value 1 between the content information 1 and each piece of content information 2 is larger than the difference value 2 between the content information 2 and each piece of content information 3, the content information 1 is determined as the content information e, and the storage name of the scanning element a is extracted based on the content information e.
In addition, the font types involved in the evaluation may be, but are not limited to, font size, font color, font special effects exemplified above.
In other embodiments, after obtaining the content information b, the mobile phone may obtain the content information b and input the content information b into an artificial intelligence model, split each piece of content information b into at least one character sequence (or called a sentence) through the artificial intelligence model, score each character sequence, and finally select the character sequence with the highest score from the character sequences corresponding to all pieces of content information b as the storage name.
In some embodiments, the artificial intelligence model described above may also train a resulting deep learning model. In some embodiments, the score may be a relevance score between the character sequence that the artificial intelligence model may recognize and the scanned content. In other embodiments, the scoring may also be determining a relative importance score between all the character sequences that are scored according to a specified parameter. For example, the character sequence score for the occurrence of the specified parameter "title" or title symbol is relatively highest. This process may be referred to as decimated naming.
In other embodiments, a theme generation model may also be configured in the handset. The topic generation model can also be a deep learning module which is trained by a large amount of data, and the topic generation module can be used for summarizing topic contents from specified text data. The above subject matter is not limited to what appears in the specified text data. Thus, after obtaining the content information b, the mobile phone can also input all the content information b into the theme generation module to obtain the corresponding theme content as a storage name. This process may also be referred to as generative naming.
In other possible examples, before the scan a is displayed in the display interface of the mobile phone, that is, before the user has not indicated to store the scan a, the mobile phone detects the user's selection operation on the designated area of the scan a, and may use the content information in the designated area as the storage name. For example, in the case where the display interface 201 of the mobile phone in fig. 3 displays the scan 1, the user may select a designated area in the display interface 201, and then the mobile phone uses the content information appearing in the designated area as the storage name of the scan 1.
As an embodiment, the above S104 may be performed by a file name extraction/generation module.
In the above, various storage name generation modes are introduced, and the generated storage names can express the content of the scanned item. The storage names are adopted to name the scanned parts, so that the scanned parts can be conveniently searched by a user, and the man-machine interaction efficiency of searching the scanned parts is improved.
Of course, it will be appreciated that in other embodiments, the handset may also co-determine the storage name in combination with one or more of the above.
For example, as shown in fig. 13, the content information b is obtained after removing the irrelevant information in the content information a. Then, the keyword recognition mode is started. In the keyword recognition mode, the mobile phone recognizes whether the keyword 3 appears in the content information b. If there is content information b having the keyword 3, a storage name is determined based on the content information b, and the flow ends. If there is no content information b having the keyword 3, it is determined that the storage name is not determined in the keyword recognition mode.
And enabling a font information judging mode under the condition that the storage name is not determined in the keyword mode. In the font information judging mode, it is judged whether or not the content information b contains content information having the largest difference between the font type and other content information. If so, a storage name is determined from the piece of content information. If not, then it is determined that the storage name is not determined in this mode.
And under the condition that the storage name is not determined in the font information judging mode, starting a preset template mode. And in a preset template mode, the mobile phone checks whether the scanned piece a corresponds to a file format. And if the scanned piece corresponds to the file format, acquiring a title display area corresponding to the file format. If there is content information b in the title display area, a corresponding storage name is determined based on the piece of content information b. If there is no content information b in the title display area or the scan a has no corresponding file format, it is determined that the storage name is not determined in this mode.
Under the condition that the storage name is not determined in the preset template mode, the name can be named by adopting a generation type name or a drawing type name, and the name is given to the scanning piece a.
In other embodiments, the handset may also extract content information for paper documents in the camera's field of view during display of the scan preview interface. For example, the OCR technology is used to extract content information for a preview frame displayed in the current preview interface and determine a layout label of the extracted content information. The process of extracting the content information and determining the layout label may refer to S102, which is not described herein.
Then, in the case where the content information f is extracted, the position area and the layout label corresponding to the content information f may be displayed in the scan preview interface. The content information f may refer to any one of the content information mentioned in the preview frame. The position information of the content information f is displayed in correspondence with the layout label. In this way, during the display of the layout tab and the location area, the mobile phone can cancel the location area and the layout tab corresponding to the display content information f and cancel the association between the layout tab and the content information f in response to the user's operation 1 on the layout tab or the corresponding location area of the content information f. In addition, the mobile phone may modify the type of the layout tag corresponding to the content information f in response to the user's operation 2 on the layout tag or the corresponding location area of the content information f.
For example, as shown in fig. 14, the mobile phone displays a scan preview interface 1401, and the scan preview interface 1401 displays a content information "number: 1234 "and a layout label 1403. After the mobile phone receives operation 1 of the user on the location area 1402, the mobile phone may display a scan preview interface 1404, where the scan preview interface 1404 does not include the location area 1402 and the layout tag 1403, and cancel the layout tag 1403 and the content information "number: 1234", is provided.
The embodiment of the application also provides an electronic device, which may include: a memory and one or more processors. The memory is coupled to the processor. The memory is for storing computer program code, the computer program code comprising computer instructions. The computer instructions, when executed by the processor, cause the electronic device to perform the steps performed by the handset in the embodiments described above. Of course, the electronic device includes, but is not limited to, the memory and the one or more processors described above.
In some embodiments, the electronic device displays a first interface, such as a scan preview interface 107 or preview interface 802, in response to a first operation by a user. When the first interface is the interface 107, as shown in fig. 1, the first operation may include clicking on the icon 102 and clicking on the control card 104. When the first interface is the preview interface 802, as shown in fig. 8, the first operation includes an operation for indicating that the camera application is enabled and an operation for selecting a document scanning function provided by the camera application.
In some embodiments, in a case where a first object is placed in a collection area of the electronic device (i.e., in a camera view of the electronic device), the electronic device displays a second interface (e.g., the presentation interface 201, the presentation interface 401, the presentation interface 803, etc.) in response to an operation of a user to instruct scanning, where the second interface includes a first scan map corresponding to the first object. The first object is a physical original to be scanned, and may be a paper document, as shown in fig. 1, where the first object is a labor contract, the corresponding first scan pattern is a scan piece 1, as shown in fig. 4, and the corresponding first scan pattern is a scan piece 3 when the first object is an additional term of the labor contract. The first object may also be a color page in a promotional sheet, a business card, a file displayed on an electronic screen (e.g., PPT, PDF, etc.), a card, a blackboard, or a whiteboard.
In some embodiments, in response to a user clicking on a first control (e.g., control 301, control 402, etc. indicating a stored control) in the second interface, the electronic device stores the first scan, identifies a plurality of pieces of first content information (e.g., content information a described above) in the first scan, and determines a first label (e.g., a layout label corresponding to content information a) corresponding to each piece of first content information in the plurality of pieces of first content information, where the first label includes any one of a header label, a footer label, a text label, a title label, a graphic label, and a single-line text label.
After storing the first scan, the electronic device displays a third interface, i.e., my document interface (interface 304, interface 405) shown in fig. 3 and 4, in response to an operation of instructing to view the stored scan, wherein the third interface includes a plurality of scan result documents including therein a first scan corresponding to a first name determined from the second content information in the first scan. For example, in fig. 3, the first name corresponding to the scanner 1 is "labor contract", the first name corresponding to the scanner 2 is "six-stage lower school unit test", the first name corresponding to the scanner 3 is "labor contract additional term", and the like, and the first names corresponding to the scanners are all derived from the content displayed in the scanner.
In addition, the second content information is content information satisfying the following condition among the plurality of pieces of first content information: the first label corresponding to the second content information is not a header label, a footer label or a graphic label, the second content information is not watermark information, file type or security level information, and the second content information does not include a preset first type keyword (e.g., keyword 1 in the foregoing embodiment).
In some embodiments, before determining the first label corresponding to each piece of the first content information in the plurality of pieces of first content information, the electronic device determines that the first scan pattern matches a preset first layout format. The first layout format may be one of the preset file formats mentioned in the foregoing embodiments. The first layout format corresponds to layout information, and may include first location information and second location information, for example. The first position information indicates a first display area of a header in the first scan (i.e., header display area), and the second position information indicates a second display area of a footer in the first scan (footer display area). The determining the first tag corresponding to each piece of the first content information in the plurality of pieces of first content information includes: when the first content information is located in the first display area, the corresponding first label is the header label; and when the first content information is positioned in the second display area, the corresponding first label is the footer label.
In some embodiments, when the first scan pattern matches the first layout format, the electronic device displays the first prompt information and the second prompt information on the second interface, where the first prompt information is used to prompt the position of the first display area in the first scan pattern, for example, may be edge information of a header display area. The second prompting information is used for prompting the position of the second display area in the first scanning image, for example, the edge information of the footer display area can be obtained. And the electronic equipment responds to the operation of the user on the first prompt information, and the display of the first prompt information is canceled. After canceling the display of the first prompt information, the electronic device may cancel the header tag of the third content information located in the first display area.
In some embodiments, the electronic device cancels the display of the second prompt in response to the user's operation on the second prompt; after canceling the display of the second prompt information, the electronic device may cancel the header tag of the fourth content information located in the second display area.
In addition, the layout information corresponding to the first layout format may further include third location information. The third position information indicates a third display area of the title in the first scan, that is, a tag display area.
If the second content information includes fifth content information in the third display area under the condition that the first scanned image is matched with the first layout format, the electronic device determines the fifth content information as the first name, or extracts a plurality of keywords from the fifth content information and combines the keywords into the first name.
In some embodiments, during the display of the second interface, the electronic device determines a fifth display area in response to a user sliding operation, the fifth display area being a contact area with the first scan during the user sliding operation. The sliding operation may be a user selection operation of a designated area of the scan piece a, or the fifth display area may be the designated area. When the second content information includes the sixth content information located in the fifth display area, the electronic device determines the sixth content information as the first name, or extracts a plurality of keywords from the sixth content information and combines the keywords into the first name.
In some embodiments, when the seventh content information is included in the second content information, the electronic device determines the seventh content information as the first name, or extracts a plurality of keywords from the seventh content information and combines the keywords into the first name; wherein the seventh content information includes a preset second category of keywords (e.g., keyword 3 in the foregoing embodiment).
In some embodiments, determining the first name from the second content information in the first scan comprises: when the eighth content information is included in the second content information, the electronic device determines the eighth content information as the first name, or extracts a plurality of keywords from the eighth content information to be combined into the first name;
wherein, the eighth content information and the ninth content information correspond to a first difference value, and the ninth content information and the tenth content information correspond to a second difference value. The first difference value between the eighth content information and each piece of ninth content information is larger than all second difference values corresponding to the ninth content information, the ninth content information is content information except the eighth content information in the second content information, the tenth content information is content information except the eighth content information and the ninth content information in the second content information, and the difference values can be determined by a preset font evaluation model based on font sizes, font colors and font special effects among different content information.
In some embodiments, the electronic device is preset with a semantic scoring model, where the semantic scoring model may be an artificial intelligence model provided in the foregoing embodiments. The electronic equipment utilizes the semantic scoring model to determine a first score corresponding to each piece of second content information, wherein the first score indicates the importance degree of the second content information, namely, the importance degree score; in other examples, the first score may also be a relevance score. The electronic equipment determines eleventh content information which is the content information with the highest corresponding first score in the second content information; the electronic device determines the eleventh content information as the first name, or extracts a plurality of keywords from the eleventh content information to be combined into the first name.
In some embodiments, the electronic device is preset with a title generation model for generalizing the meaning, which may also be referred to as a theme generation model. Determining the first name according to the second content information in the first scan comprises: the electronic equipment processes the second content information by using the title generation model, and determines a first meaning corresponding to the second content information, namely, a theme content. The electronic device determines the first meaning as the first name.
The embodiment of the application also provides a chip system, which can be applied to the terminal equipment in the previous embodiment. As shown in fig. 15, the system-on-chip includes at least one processor 2201 and at least one interface circuit 2202. The processor 2201 may be a processor in an electronic device as described above. The processor 2201 and the interface circuit 2202 may be interconnected by wires. The processor 2201 may receive and execute computer instructions from the memory of the electronic device described above through the interface circuit 2202. The computer instructions, when executed by the processor 2201, cause the electronic device to perform the steps performed by the handset in the embodiments described above. Of course, the chip system may also include other discrete devices, which are not specifically limited in this embodiment of the present application.
In some embodiments, it will be clearly understood by those skilled in the art from the foregoing description of the embodiments, for convenience and brevity of description, only the division of the above functional modules is illustrated, and in practical application, the above functional allocation may be implemented by different functional modules, that is, the internal structure of the apparatus is divided into different functional modules to implement all or part of the functions described above. The specific working processes of the above-described systems, devices and units may refer to the corresponding processes in the foregoing method embodiments, which are not described herein.
The functional units in the embodiments of the present application may be integrated in one processing unit, or each unit may exist alone physically, or two or more units may be integrated in one unit. The integrated units may be implemented in hardware or in software functional units.
The integrated units, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the embodiments of the present application may be essentially or a part contributing to the prior art or all or part of the technical solution may be embodied in the form of a software product stored in a storage medium, including several instructions to cause a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor to perform all or part of the steps of the methods described in the embodiments of the present application. And the aforementioned storage medium includes: flash memory, removable hard disk, read-only memory, random access memory, magnetic or optical disk, and the like.
The foregoing is merely a specific implementation of the embodiments of the present application, but the protection scope of the embodiments of the present application is not limited thereto, and any changes or substitutions within the technical scope disclosed in the embodiments of the present application should be covered by the protection scope of the embodiments of the present application. Therefore, the protection scope of the embodiments of the present application shall be subject to the protection scope of the claims.

Claims (12)

1. A data processing method, applied to an electronic device, the method comprising:
the electronic equipment responds to a first operation of a user and displays a first interface, wherein the first interface is a scanning preview interface;
under the condition that a first object is placed in an acquisition area of the electronic equipment, the electronic equipment responds to an operation of user indication scanning, and a second interface is displayed, wherein the second interface comprises a first scanning image corresponding to the first object;
responding to clicking operation of a first control in the second interface by a user, storing the first scanning image by the electronic equipment, identifying a plurality of pieces of first content information in the first scanning image, and determining a first label corresponding to each piece of first content information in the plurality of pieces of first content information, wherein the first label comprises any one of a header label, a page footer label, a text label, a title label, a graphic label and a single-row text label;
After storing the first scan, the electronic device displays a third interface in response to an operation of indicating to view the stored scan, wherein the third interface comprises a plurality of scan result documents, the plurality of scan result documents comprise the first scan, the first scan corresponds to a first name, and the first name is determined according to second content information in the first scan;
wherein the second content information is content information satisfying the following condition among the plurality of pieces of first content information: the first label corresponding to the second content information is not a header label, a footer label and a graphic label, the second content information is not watermark information, file type and security level information, and the second content information does not contain preset first keywords.
2. The method of claim 1, wherein prior to determining the first tag for each of the plurality of pieces of first content information, the method further comprises:
the electronic equipment determines that the first scanning image is matched with a preset first layout format; the first layout format comprises first position information and second position information, wherein the first position information indicates a first display area of a header in the first scanning diagram, and the second position information indicates a second display area of a footer in the first scanning diagram;
The determining the first tag corresponding to each piece of the first content information in the plurality of pieces of first content information includes:
when the first content information is located in the first display area, the corresponding first label is the header label;
and when the first content information is positioned in the second display area, the corresponding first label is the footer label.
3. The method of claim 2, wherein after displaying the second interface, the method further comprises:
when the first scanned image is matched with the first layout format, the electronic equipment displays first prompt information and second prompt information on the second interface, wherein the first prompt information is used for prompting the position of the first display area in the first scanned image, and the second prompt information is used for prompting the position of the second display area in the first scanned image;
the electronic equipment responds to the operation of a user on the first prompt information, and the display of the first prompt information is canceled;
after the first prompt information is canceled from being displayed, the determining a first tag corresponding to each piece of first content information in the plurality of pieces of first content information includes: canceling the header tag of third content information, the third content information being the first content information located in the first display area;
Or, the electronic equipment responds to the operation of the user on the second prompt information, and the display of the second prompt information is canceled;
after the second prompt information is canceled from being displayed, the determining the first tag corresponding to each piece of first content information in the plurality of pieces of first content information includes: and canceling the header tag of fourth content information, wherein the fourth content information is the first content information positioned in the second display area.
4. The method of claim 2, wherein the first layout format further corresponds to third location information indicating a third display area of a title in the first scan, and wherein determining the first name based on the second content information in the first scan comprises:
and if the second content information comprises fifth content information positioned in the third display area under the condition that the first scanning image is matched with the first layout format, the electronic equipment determines the fifth content information as the first name, or extracts a plurality of keywords from the fifth content information and combines the keywords into the first name.
5. The method of claim 1, wherein during displaying the second interface, the method further comprises:
the electronic device determining a fifth display area in response to a sliding operation of a user, the fifth display area being a contact area with the first scan pattern during the sliding operation of the user;
determining the first name according to the second content information in the first scan comprises:
when the second content information includes sixth content information located in the fifth display area, the electronic device determines the sixth content information as the first name, or extracts a plurality of keywords from the sixth content information and combines the keywords into the first name.
6. The method of claim 1, wherein determining the first name from the second content information in the first scan comprises:
when the second content information comprises seventh content information, the electronic equipment determines the seventh content information as the first name, or extracts a plurality of keywords from the seventh content information to be combined into the first name; wherein the seventh content information includes a preset second category keyword.
7. The method of claim 1, wherein determining the first name from the second content information in the first scan comprises:
when the second content information comprises eighth content information, the electronic equipment determines the eighth content information as the first name, or extracts a plurality of keywords from the eighth content information to be combined into the first name;
a first difference value is corresponding between the eighth content information and the ninth content information, a second difference value is corresponding between the ninth content information and the tenth content information, the first difference value is larger than the second difference value, the ninth content information is content information except for the eighth content information in the second content information, the tenth content information is content information except for the eighth content information and the ninth content information in the second content information, and the difference value can be determined by a preset font evaluation model based on font sizes, font colors and font special effects among different content information.
8. The method of claim 1, wherein the electronic device is pre-configured with a semantic scoring model, wherein determining the first name based on the second content information in the first scan comprises:
The electronic equipment determines a first score corresponding to each piece of second content information by using the semantic score model, wherein the first score indicates the importance degree of the second content information;
the electronic equipment determines eleventh content information, wherein the eleventh content information is content information with the highest first score corresponding to the second content information;
the electronic device determines the eleventh content information as the first name, or extracts a plurality of keywords from the eleventh content information to be combined into the first name.
9. The method of claim 1, wherein the electronic device is preset with a title generation model for generalizing text, and wherein determining the first name from the second content information in the first scan comprises:
the electronic equipment processes the second content information by using the title generation model and determines a first meaning corresponding to the second content information;
the electronic device determines the first meaning as the first name.
10. An electronic device comprising one or more processors and memory; the memory being coupled to a processor, the memory being for storing computer program code comprising computer instructions which, when executed by one or more processors, are for performing the method of any of claims 1-9.
11. A computer storage medium comprising computer instructions which, when run on an electronic device, cause the electronic device to perform the method of any of claims 1-9.
12. A computer program product, characterized in that the computer program product comprises a computer program which, when run on a computer, causes the computer to perform the method according to any of claims 1-9.
CN202211214661.5A 2022-09-30 2022-09-30 Data processing method and electronic equipment Pending CN117812187A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211214661.5A CN117812187A (en) 2022-09-30 2022-09-30 Data processing method and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211214661.5A CN117812187A (en) 2022-09-30 2022-09-30 Data processing method and electronic equipment

Publications (1)

Publication Number Publication Date
CN117812187A true CN117812187A (en) 2024-04-02

Family

ID=90418961

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211214661.5A Pending CN117812187A (en) 2022-09-30 2022-09-30 Data processing method and electronic equipment

Country Status (1)

Country Link
CN (1) CN117812187A (en)

Similar Documents

Publication Publication Date Title
CN110196675B (en) Method for adding annotations and electronic equipment
WO2020238356A1 (en) Interface display method and apparatus, terminal, and storage medium
CN111465918B (en) Method for displaying service information in preview interface and electronic equipment
WO2021258797A1 (en) Image information input method, electronic device, and computer readable storage medium
CN114422640B (en) Equipment recommendation method and electronic equipment
US20220050975A1 (en) Content Translation Method and Terminal
WO2019105457A1 (en) Image processing method, computer device and computer readable storage medium
CN116302227A (en) Method for combining multiple applications and simultaneously starting multiple applications and electronic equipment
US11263447B2 (en) Information processing method, information processing device, mobile terminal, and storage medium
KR20080114389A (en) Method and apparatus for printing data of predetermined format using bluetooth communication, method for storing template data
CN112001312A (en) Document splicing method, device and storage medium
WO2021218365A1 (en) Annotation method and electronic device
US9854132B2 (en) Image processing apparatus, data registration method, and data registration program
CN114943976B (en) Model generation method and device, electronic equipment and storage medium
CN114117269B (en) Memo information collection method and device, electronic equipment and storage medium
CN117812187A (en) Data processing method and electronic equipment
CN116012570A (en) Method, equipment and system for identifying text information in image
CN114594882A (en) Feedback method, device and system
WO2023011143A1 (en) Card information management method and electronic device
CN117851617A (en) Display method, electronic device, storage medium, and program product
CN115879436B (en) Electronic book quality inspection method
CN116050390A (en) Text processing method and electronic equipment
CN113742565A (en) Content classification method, device and system
KR20140081516A (en) Online Bank Document System
CN114528468A (en) Display method and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination