WO2021169400A1 - 基于人工智能的命名实体识别方法、装置及电子设备 - Google Patents
基于人工智能的命名实体识别方法、装置及电子设备 Download PDFInfo
- Publication number
- WO2021169400A1 WO2021169400A1 PCT/CN2020/127737 CN2020127737W WO2021169400A1 WO 2021169400 A1 WO2021169400 A1 WO 2021169400A1 CN 2020127737 W CN2020127737 W CN 2020127737W WO 2021169400 A1 WO2021169400 A1 WO 2021169400A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- text
- candidate entity
- recognized
- named entity
- word
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
- G06F40/289—Phrasal analysis, e.g. finite state techniques or chunking
- G06F40/295—Named entity recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/36—Creation of semantic tools, e.g. ontology or thesauri
- G06F16/367—Ontology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/166—Editing, e.g. inserting or deleting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computing arrangements using knowledge-based models
- G06N5/02—Knowledge representation; Symbolic representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Definitions
- This application relates to artificial intelligence technology, in particular to a named entity recognition method, device, electronic equipment, and computer-readable storage medium based on artificial intelligence.
- AI Artificial Intelligence
- NLP Nature Language Processing
- Named entity recognition is an important research branch of natural language processing, which aims to locate and classify named entities in texts into predefined categories, such as persons, organizations, locations, or numbers.
- a certain part of a named entity may also be a named entity, that is, there are multiple levels of nesting. Therefore, there is a need to identify multiple levels of nested named entities.
- the embodiment of the present application provides an artificial intelligence-based named entity recognition method, including:
- the embodiment of the present application provides a named entity recognition device based on artificial intelligence, including:
- the vector conversion module is configured to perform vector conversion processing on the text element in the text to be recognized to obtain the text representation of the text element;
- a construction module configured to construct candidate entity words according to the text elements whose total number does not exceed the element number threshold in the text to be recognized;
- An integration module configured to perform integration processing on the text representation corresponding to the text element in the candidate entity word to obtain the text representation of the candidate entity word
- the classification module is configured to classify the text representation of the candidate entity word to determine the category to which the candidate entity word belongs among the non-named entity category and multiple named entity categories.
- An embodiment of the application provides an electronic device, including:
- Memory used to store executable instructions
- the processor is configured to implement the artificial intelligence-based named entity recognition method provided in the embodiment of the present application when executing the executable instructions stored in the memory.
- the embodiment of the present application provides a computer-readable storage medium storing executable instructions for causing a processor to execute, to implement the artificial intelligence-based named entity recognition method provided by the embodiment of the present application.
- FIG. 1 is a schematic diagram of an architecture of a named entity recognition system based on artificial intelligence provided by an embodiment of the present application
- FIG. 2 is a schematic diagram of an architecture of a server provided by an embodiment of the present application.
- FIG. 3 is a schematic diagram of an architecture of an artificial intelligence-based named entity recognition device provided by an embodiment of the present application
- FIG. 4A is a schematic flowchart of a named entity recognition method based on artificial intelligence provided by an embodiment of the present application
- 4B is a schematic flowchart of a named entity recognition method based on artificial intelligence provided by an embodiment of the present application
- FIG. 4C is a schematic flowchart of a named entity recognition method based on artificial intelligence provided by an embodiment of the present application.
- FIG. 4D is a schematic flowchart of a named entity recognition method based on artificial intelligence provided by an embodiment of the present application.
- FIG. 5 is a schematic diagram of an architecture for using a recurrent neural network model for named entity recognition according to an embodiment of the present application
- FIG. 6 is a schematic diagram of an architecture for using a convolutional neural network model for named entity recognition according to an embodiment of the present application
- Fig. 7 is a schematic flowchart of a question and answer scenario provided by an embodiment of the present application.
- first ⁇ second involved only distinguishes similar objects, and does not represent a specific order for the objects. Understandably, “first ⁇ second” can be used if permitted. The specific order or sequence is exchanged, so that the embodiments of the present application described herein can be implemented in a sequence other than those illustrated or described herein.
- plural refers to at least two.
- Natural language processing It is an important direction of artificial intelligence, studying language problems in human-to-human communication and human-computer communication.
- natural language processing we mainly develop models that express language capabilities (Linguistic Competence) and language applications (Linguistic Performance), establish a computing framework to implement such language models, and propose corresponding methods to continuously improve such language models.
- the language model designs various practical systems and discusses the evaluation techniques of these practical systems.
- Convolutional Neural Network (CNN, Convolutional Neural Network) model a feedforward neural network model in which artificial neurons can respond to surrounding units.
- the convolutional neural network model usually includes a convolutional layer and a pooling layer.
- Recurrent Neural Network (RNN, Recurrent Neural Network) model a type of neural network model used to process sequence data.
- the internal state of this model can show dynamic timing behavior, and can use internal memory to process variable-length input sequences .
- the cyclic update processing can be implemented through the RNN model.
- Named Entity An entity with a specific meaning in the text. According to different actual application scenarios, named entities can have different categories and regulations. For example, named entities can include persons, organizations, locations, time expressions, quantities, currency values, and percentages.
- Multi-category named entities Named entities that belong to multiple named entity categories at the same time.
- Multi-level nested named entities A certain part of a named entity is also a named entity, which is called multi-level nesting. For example, in the text "Glorious People's Liberation Army of a certain country”, “People's Liberation Army of a certain country” is a named entity of the "army” category, and "a country” nested in the named entity “People's Liberation Army of a certain country” is also a named entity of the category “country” .
- Sequence Tagging usually refers to the process of tagging each element in the sequence with a certain tag in the tag set for a linear input sequence.
- Named Entity Recognition aims to locate and classify named entities in the text into predefined categories, such as persons, organizations, and locations.
- NER Named Entity Recognition
- the results of NER can be applied to application scenarios such as information extraction, question answering systems, syntax analysis, and machine translation.
- Nested NER Traditional NER can only roughly identify text with a flat structure from the text. Unlike traditional NER, Nested NER’s task goal is to identify multiple categories of named entities and multiple levels of nesting from the text. Named entity.
- Text representation is a high-level cognitive abstract entity produced in the process of human cognition. In natural language processing, text needs to be converted into a data type that can be processed by a computer, that is, it is converted into a vector form of text representation.
- Text element In the embodiment of the present application, the text element can be a character or a word, that is, the number of characters included in the text element is not limited.
- the related technology mainly provides the Nested NE BILOU coding and decoding scheme to realize it.
- the scheme mainly performs sequence labeling of tokens and performs coding and decoding processing based on the sequence-to-sequence model structure.
- a labeling coding method similar to BILOU is adopted.
- this scheme allows multi-layer token labeling.
- the solution formulates certain rules to process these multi-layer label results, such as nearest matching.
- the following coding example table is provided:
- B means the starting position of the named entity
- I means the middle of the named entity
- L means the end of the named entity
- O means not belonging to the named entity
- U means a separate named entity.
- ORG means an organization
- GPE means a geopolitical entity.
- the embodiments of the present application provide an artificial intelligence-based named entity recognition method, device, electronic device, and computer-readable storage medium, which can improve the efficiency and accuracy of named entity recognition.
- the electronic device provided by the embodiment of the application may be a server, such as a server deployed in the cloud, and provide users with remote named entity recognition based on the acquired text to be recognized. Function; it can also be a terminal device, such as a question and answer device, which expands the knowledge graph based on the named entity recognized by the named entity, and realizes intelligent question answering based on the knowledge graph; it can even be a handheld terminal and other devices.
- a server such as a server deployed in the cloud, and provide users with remote named entity recognition based on the acquired text to be recognized.
- Function can also be a terminal device, such as a question and answer device, which expands the knowledge graph based on the named entity recognized by the named entity, and realizes intelligent question answering based on the knowledge graph; it can even be a handheld terminal and other devices.
- FIG. 1 is a schematic diagram of an optional architecture of an artificial intelligence-based named entity recognition system 100 provided by an embodiment of the present application.
- the terminal device 400 is connected through the network 300
- the server 200 and the network 300 may be a wide area network or a local area network, or a combination of the two.
- the artificial intelligence-based named entity recognition method provided in the embodiments of the present application may be implemented by the terminal device.
- the terminal device 400 obtains the text to be recognized entered or automatically selected by the user, and determines the text representation of the text element in the text to be recognized. Then, based on the text elements in the text to be recognized, the terminal device 400 constructs multiple candidate entity words according to the set element number threshold, and for each candidate entity word obtained, integrates the text representations corresponding to the text elements in the candidate entity words Processing, and classifying the obtained text representation of the candidate entity word, so as to obtain the category to which the candidate entity word belongs.
- the artificial intelligence-based named entity recognition method provided in the embodiments of the present application may also be implemented by the server.
- the server 200 obtains the text to be recognized from the database, and constructs multiple candidate entity words according to a set threshold of the number of elements. For each candidate entity word, the text representation of the candidate entity word is classified, so as to obtain the category to which the candidate entity word belongs.
- the artificial intelligence-based named entity recognition method provided in the embodiments of the present application may also be implemented by a terminal device and a server in cooperation.
- the server 200 obtains the to-be-recognized text sent by the terminal device 400, and after a series of processing, sends the identified candidate entity words belonging to the named entity category to the terminal device 400 so that the user of the terminal device 400 can be informed. That is, the server 200 is configured to execute the Nested NER task, and the terminal device 400 is configured to collect the processing object of the Nested NER task (that is, the text to be recognized), and present the execution result of the Nested NER task.
- the terminal device 400 can display various results of the named entity recognition process in the graphical interface 410, such as candidate entity words belonging to the category of named entities.
- candidate entity words belonging to the category of named entities In FIG. 1, the text "Glorious People's Liberation Army of a certain country" to be recognized is taken as an example.
- Candidate entity words belonging to the category of named entities in the text to be recognized are listed, including "a certain country”, “a certain countryman” and "a certain country's People's Liberation Army”.
- the results of named entity recognition can be applied to various application scenarios in the NLP field, such as the scenarios of summary determination, object recommendation, text classification, and question answering system shown in Figure 1, and the scenarios of information extraction, syntax analysis, and machine translation. An example application of named entities will be explained later.
- the terminal device may run a computer program to implement the artificial intelligence-based named entity recognition method provided in the embodiments of the present application.
- the computer program may be a native program or a software module in the operating system; it may be a local ( Native) application (APP, Application), that is, a program that needs to be installed in the operating system to run; it can also be a small program, that is, a program that only needs to be downloaded to the browser environment to run; it can also be embedded in any Small programs in the APP.
- APP Native
- the above-mentioned computer program may be any form of application, module or plug-in.
- Cloud technology refers to a kind of hosting that unifies a series of resources such as hardware, software, and network in a wide area network or a local area network to realize the calculation, storage, processing and sharing of data.
- cloud technology is also a general term for network technology, information technology, integration technology, management platform technology, and application technology based on the application of cloud computing business models. It can form a resource pool, which can be used on demand and is flexible and convenient. Cloud computing technology will become an important support. The background service of the technical network system requires a lot of computing and storage resources.
- the above-mentioned server may be an independent physical server, or a server cluster or a distributed system composed of multiple physical servers, or it may provide cloud services, cloud databases, cloud computing, cloud functions, and cloud storage.
- Cloud servers for basic cloud computing services such as network services, cloud communications, middleware services, domain name services, security services, CDN, and big data and artificial intelligence platforms.
- cloud services can be named entity recognition services for terminal devices transfer.
- the terminal device can be a smart phone, a tablet computer, a notebook computer, a desktop computer, a smart watch, a smart TV, etc., but it is not limited to this.
- the terminal device and the server can be directly or indirectly connected through wired or wireless communication, which is not limited in the embodiment of the present application.
- FIG. 2 is a schematic structural diagram of a server 200 (for example, the server 200 shown in FIG. 1) provided by an embodiment of the present application.
- the server 200 shown in FIG. 2 includes: at least one processor 210, a memory 240, and At least one network interface 220.
- the various components in the server 200 are coupled together through the bus system 230.
- the bus system 230 is used to implement connection and communication between these components.
- the bus system 230 also includes a power bus, a control bus, and a status signal bus.
- various buses are marked as the bus system 230 in FIG. 2.
- the processor 210 may be an integrated circuit chip with signal processing capabilities, such as a general-purpose processor, a digital signal processor (DSP, Digital Signal Processor), or other programmable logic devices, discrete gates or transistor logic devices, or discrete hardware Components, etc., where the general-purpose processor may be a microprocessor or any conventional processor.
- DSP Digital Signal Processor
- the memory 240 may be removable, non-removable, or a combination thereof.
- Exemplary hardware devices include solid-state memory, hard disk drives, optical disk drives, and so on.
- the memory 240 optionally includes one or more storage devices that are physically remote from the processor 210.
- the memory 240 includes volatile memory or nonvolatile memory, and may also include both volatile and nonvolatile memory.
- the non-volatile memory may be a read only memory (ROM, Read Only Memory), and the volatile memory may be a random access memory (RAM, Random Access Memory).
- ROM read only memory
- RAM random access memory
- the memory 240 described in the embodiment of the present application is intended to include any suitable type of memory.
- the electronic device may also include a user interface.
- the user interface may include one or more output devices that enable the presentation of media content, including One or more speakers and/or one or more visual display screens.
- the user interface may also include one or more input devices, including user interface components that facilitate user input, such as a keyboard, a mouse, a microphone, a touch screen display, a camera, and other input buttons and controls.
- the memory 240 can store data to support various operations. Examples of these data include programs, modules, and data structures, or a subset or superset thereof, as illustrated below.
- the operating system 241 includes system programs for processing various basic system services and performing hardware-related tasks, such as a framework layer, a core library layer, a driver layer, etc., for implementing various basic services and processing hardware-based tasks;
- the network communication module 242 is used to reach other computing devices via one or more (wired or wireless) network interfaces 220.
- Exemplary network interfaces 220 include: Bluetooth, Wireless Compatibility Authentication (WiFi), and Universal Serial Bus ( USB, Universal Serial Bus), etc.
- the memory 240 may further include a presentation module and an input processing module.
- the presentation module is used to enable presentation of information (for example, a user interface for operating peripheral devices and displaying content and information) via one or more output devices (for example, a display screen, a speaker, etc.) associated with a user interface;
- the input processing module is used to detect one or more user inputs or interactions from one of the one or more input devices and translate the detected inputs or interactions.
- the artificial intelligence-based named entity recognition device provided by the embodiments of the present application can be implemented in software.
- FIG. 2 shows the artificial intelligence-based named entity recognition device 243 stored in the memory 240, which may be Software in the form of programs and plug-ins, including the following software modules: vector conversion module 2431, building module 2432, integration module 2433, and classification module 2434. These modules are logical, so they can be combined or further based on the implemented functions. Split. The function of each module will be explained below.
- Figure 3 is a schematic diagram of the architecture of an artificial intelligence-based named entity recognition device 243 provided in an embodiment of the present application, showing the process of implementing named entity recognition through a series of modules.
- Figure 4A is an embodiment of the present application. The provided flow chart of the artificial intelligence-based named entity recognition method will describe the steps shown in FIG. 4A in conjunction with FIG. 3.
- step 101 the text element in the text to be recognized is subjected to vector conversion processing to obtain the text representation of the text element.
- the text to be recognized is obtained, where the text to be recognized can be entered by the user, such as by voice, handwriting or other means, or automatically selected, such as electronic
- the device can use the locally stored text as the text to be recognized.
- multiple text elements in the text to be recognized are determined, where the type of the text element can be a character or a word.
- vector conversion processing is performed on the text element, that is, embedding processing is performed to map the text element to the vector space to obtain a text representation in the form of a vector, which is convenient for subsequent processing.
- the method before step 101, further includes: performing any one of the following operations to obtain text elements in the text to be recognized: treating each word in the text to be recognized as a text element; performing word segmentation processing on the text to be recognized , The word obtained by word segmentation processing is used as a text element.
- the embodiments of the present application provide different determination methods.
- the type of text element to be determined is a word
- each word in the text to be recognized can be used as a text element.
- the number of text elements determined in this way is large, and the subsequent calculations are relatively large.
- the accuracy of entity recognition is high.
- word segmentation processing can be performed on the text to be recognized, and each word obtained by the word segmentation processing can be used as a text element.
- the word segmentation processing can be implemented through a language technology platform (LTP, Language Technology Platform) tool or other tools, which is not limited.
- LTP Language Technology Platform
- the number of text elements obtained in this way is small, that is, the amount of subsequent calculations is small, and the efficiency of named entity recognition is high. Either of the above two methods can be selected according to the requirements in the actual application scenario.
- the text to be recognized can be segmented first, for example, using punctuation marks (such as commas and periods, etc.) in the text to be recognized as the segmentation position for segmentation. Obtain multiple sentences in the text to be recognized, and then perform word segmentation processing on each sentence to determine the text element in each sentence.
- the above-mentioned segmentation processing method is suitable for the case where the text to be recognized includes multiple sentences.
- step 102 candidate entity words are constructed according to the text elements whose total number does not exceed the element number threshold in the text to be recognized.
- each candidate entity word obtained is a part of the text to be recognized. For example, if the text to be recognized is "Glorious People's Liberation Army of a certain country", then a certain candidate entity word "People's Liberation Army of a certain country” is a part of the text to be recognized.
- the element number threshold can be manually set. For example, if it is set to 7, the named entity with the most text elements included in the database can also be obtained, and the number of text elements included in the named entity can be determined as the element number threshold.
- step 103 the text representation corresponding to the text element in the candidate entity word is integrated to obtain the text representation of the candidate entity word.
- step 102 multiple candidate entity words may be obtained.
- the text representations corresponding to all text elements in the candidate entity words are integrated to obtain the text representation of the candidate entity word.
- the embodiment of the present application does not limit the way of integration processing. For example, it may be weighted processing (such as weighted summation), etc., depending on the way of constructing candidate entity words.
- step 104 the text representation of the candidate entity word is classified, so as to determine the category to which the candidate entity word belongs among the non-named entity category and multiple named entity categories.
- the text representation of the candidate entity word can accurately and effectively represent the semantics of the candidate entity word. Therefore, the text representation of the candidate entity word can be classified to determine the category of the candidate entity word. For example, the text representation of the candidate entity word can be fully connected, and then the text representation after the fully connected process can be mapped to the probability corresponding to the non-named entity category and the probability of one-to-one correspondence with multiple named entity categories. According to these The probability determines the category to which the candidate entity word belongs. It is worth noting that in the embodiments of the present application, the non-named entity is also regarded as a category, and is treated the same as the named entity category.
- the candidate entity word belonging to the named entity category (herein refers to any named entity category) in the text to be recognized can be applied to natural language processing application scenarios, including but not limited to Figure 3
- the application scenarios of summary determination, object recommendation, text classification and question answering are shown.
- the method further includes: segmenting the text to be recognized to obtain multiple sentences; determining the candidate entity words belonging to any named entity category and whose appearance frequency meets the frequency condition as abstract keywords; The number of summary keywords included in the sentence determines the score of the sentence; the sentence whose score meets the scoring condition is determined as the text summary of the text to be recognized.
- the text summary of the text to be recognized can be determined according to the result of named entity recognition.
- the text to be recognized can be a paper, news, or review article, etc., which is not limited.
- punctuation marks such as commas and periods
- the segmentation processing operation can also be performed before step 101. The application embodiment does not limit this.
- the candidate entity words belonging to any named entity category in the text to be recognized are named as keywords, and the frequency of occurrence of the keywords in the text to be recognized is determined.
- the keyword is determined as an abstract keyword.
- the K keywords with the highest frequency of occurrence are determined as summary keywords, where K is an integer greater than 0.
- the keywords corresponding to the frequency of occurrence exceeding the frequency threshold can also be determined as summary keywords.
- the frequency threshold can be determined according to Actual application scenario setting.
- the number of summary keywords included in the sentence is determined, and the score of the sentence is determined according to the number.
- the number of summary keywords included in the sentence can be directly used as the score of the sentence, or the number of summary keywords included in the sentence can be divided by the total number of text elements included in the sentence to obtain the score of the sentence.
- the sentence whose score meets the scoring conditions is determined as the text summary of the text to be recognized.
- the L sentences with the highest score can be determined as the text summary of the text to be recognized, where L is an integer greater than 0, or the score
- the sentence that exceeds the scoring threshold is determined as a text summary of the text to be recognized.
- step 104 after step 104, it further includes: when the text to be recognized is used to represent the object to be recommended, determining the candidate entity word belonging to any named entity category as the keyword of the object to be recommended; obtaining the key to the user portrait The keyword coincidence degree between the keyword of the user portrait and the keyword of the object to be recommended is determined; when the degree of keyword coincidence exceeds the first coincidence degree threshold, the operation of recommending the object to be recommended is performed.
- the text to be recognized is used to indicate the object to be recommended, and the embodiment of this application does not limit the type of the object to be recommended.
- the object to be recommended can be a product, and the text to be recognized is the product description of the product; the object to be recommended can also be It is a movie, and the text to be recognized is a summary description of the movie.
- targeted smart object recommendation can be realized.
- the candidate entity words belonging to any named entity category in the text to be recognized are determined as the keywords of the object to be recommended, and at the same time, user portrait keywords are obtained, such as obtaining user portraits of all registered users in an application Key words. Then, determine the degree of keyword overlap between the user portrait keywords of each user and the keywords to be recommended.
- the user portrait keywords can be set by the user, or they can be the key to the user's historical browsing records Word statistics.
- the intersection and union between the user profile keywords and the keywords to be recommended can be determined, and the first number of keywords included in the intersection and the second number of keywords included in the union can be determined Divide, get the keyword coincidence degree.
- the keywords in the text to be recognized include “love”, “artistic” and “story”
- the keywords of the user portrait include “love”, “science fiction” and For “comedy”
- the keyword overlap is 1/5.
- the first coincidence degree threshold is set to determine whether to recommend, for example, when the keyword coincidence degree between the user portrait keyword of a certain user and the keyword to be recommended exceeds the first coincidence degree threshold (For example, 80%), the operation of recommending the object to be recommended to the user is performed.
- the embodiment of the application does not limit the specific method of recommendation, for example, it may be email recommendation, SMS recommendation, or front-end pop-up recommendation.
- the recommended object is made more in line with the user's portrait, that is, in line with the user's interest, the accuracy of the recommendation is increased, and the actual utilization of the computing resources consumed by the electronic device when making the recommendation is improved.
- the method further includes: determining a candidate entity word belonging to any named entity category as a keyword; determining the degree of keyword overlap between the first text to be recognized and the second text to be recognized; When the keyword coincidence degree exceeds the second coincidence degree threshold, the first to-be-recognized text and the second to-be-recognized text are classified into the same text category.
- the result of named entity recognition can also be used in text classification application scenarios.
- the candidate entity words belonging to any named entity category in the text to be recognized are determined as keywords
- the first text to be recognized and the second text to be recognized are determined
- the degree of overlap of keywords between texts are determined.
- the intersection and union between the keywords of the first text to be recognized and the keywords of the second text to be recognized can also be determined according to the above method, and the number of keywords included in the intersection Divide by the number of keywords included in the union to obtain the keyword overlap between the first text to be recognized and the second text to be recognized.
- a second coincidence degree threshold is also set.
- the keyword coincidence degree exceeds the second coincidence degree threshold (such as 80%)
- Texts classified into the same text category can be used for similar text recommendation. For example, when the query target of a text query request is the first text to be recognized, the first text to be recognized and the first text in the same text category as the first text to be recognized are returned. Second, the text to be recognized in response to the text query request.
- the keyword coincidence degree is regarded as the similarity between the texts, and the text classification is carried out to improve the accuracy of the classification.
- the classified texts can be used for similar text recommendation, that is, the texts belonging to the same text category Co-recommend to users, which can effectively improve user experience.
- step 104 further includes: adding candidate entity words belonging to any named entity category to the knowledge graph; wherein the knowledge graph is used to respond to category queries for candidate entity words belonging to any named entity category ask.
- the result of named entity recognition can be used for new word discovery.
- the candidate entity words belonging to any named entity category in the text to be recognized are determined as keywords, and the keywords and the category to which the keywords belong are added to the knowledge graph together.
- the knowledge graph can be used to respond to the category query request for the candidate entity words belonging to the named entity category.
- a certain candidate entity word added to the knowledge graph is "a country”, and its category is named entity category "country”.
- a category query request including "a country” is received, it can be searched in the knowledge graph , Respond to the query request of the category according to the "country” category obtained in the query, that is, respond.
- step 104 after step 104, it further includes: determining a candidate entity word belonging to any named entity category as a keyword; performing syntactic analysis on the text to be recognized to obtain subject keywords, relational words, and words in the text to be recognized Object keywords; among them, relative words are used to express the relationship between subject keywords and object keywords; construct triples according to subject keywords, relative words and object keywords, and add the triples to the knowledge graph; , The knowledge graph is used to respond to object query requests including subject keywords and relative words.
- the relationship between keywords can also be added to the knowledge graph.
- the text to be recognized can be syntactically analyzed.
- syntactic analysis can be performed through LTP tools or other tools.
- the subject keywords, relative words, and object keywords in the text to be recognized are obtained, where the relative words are used to express the relationship between the subject keywords and the object keywords.
- the text to be recognized is "Zhang San borrows money from Li Si”
- the keywords include "Zhang San" and "Li Si”
- the subject keyword is "Zhang San” and the relative word is "Zhang San”.
- “Borrow money” the object keyword is "Li Si”.
- the knowledge graph can be used to respond to object query requests that include subject keywords and relational words. For example, when the object query request is used to query the object of Zhang San's loan, it can be searched in the knowledge graph and the result is "Li Si". .
- the object query request is used to query the object of Zhang San's loan, it can be searched in the knowledge graph and the result is "Li Si". .
- the embodiment of the application constructs candidate entity words whose total number of text elements does not exceed the element number threshold, which can effectively cope with the situation where named entities have multiple levels of nesting, greatly improving the efficiency and flexibility of named entity recognition. sex.
- FIG. 4B is an optional flowchart of the artificial intelligence-based named entity recognition method provided by an embodiment of the present application.
- Step 102 shown in FIG. 4A can be implemented through step 201 to step 202. , Will be described in conjunction with each step.
- step 201 traversal processing is performed on the text elements in the text to be recognized.
- the traversal processing and the scanning loop can be combined to obtain the candidate entity words.
- the traversal processing is performed on the text elements in the text to be recognized.
- the embodiment of the present application does not limit the sequence of the traversal processing.
- the sequence of the traversal processing can be from the first text element to the last text element in the text to be recognized.
- the order can also be the order from the last text element to the first text element in the text to be recognized.
- step 202 for the traversed text elements, a scanning loop including multiple scanning processes is executed, and the traversed text elements and the text elements obtained from each scanning process are combined into candidate entity words; wherein, the candidate entity words include The total number of text elements does not exceed the element number threshold.
- a scan cycle is executed, and the scan cycle includes multiple scan processing.
- the traversed text elements and the text elements obtained from the scanning process are combined into candidate entity words.
- a candidate entity word can be obtained, where the candidate entity word The total number of included text elements does not exceed the element number threshold.
- the execution of the scanning cycle including multiple scanning processes described above can be implemented in such a way: according to the number of scanning processes that have been executed in the scanning cycle, the number of scans to be increased or decreased simultaneously is determined, according to The number of scans executes the scanning process starting from the traversed text elements, and combines the traversed text elements and the text elements obtained by the scanning process into candidate entity words, until the initial zero scan number increases to the scan number threshold, or the initial The number of scans, which is the threshold of the number of scans, is reduced to zero; where the threshold of the number of scans is the result obtained by subtracting one from the threshold of the number of elements.
- the description is divided into two cases.
- the first case is that the number of scans is initially 0, and the number of scans increases synchronously with the number of scans that have been executed in the scan cycle. For example, each time the number of scans that have been executed increases, the number of scans is increased by 1 accordingly. .
- the second cycle the number of scanning processes that have been executed is 0, the number of scans is 0, and the scanning process starting from "light” is executed according to the number of scans.
- the second case is that the number of scans is initially the threshold of the number of scans, and the number of scans decreases synchronously with the number of scans that have been executed in the scan cycle. For example, each time the number of scans that have been executed increases, the number of scans is reduced by 1 accordingly. operate. Take the text to be recognized as "Glorious People's Liberation Army", and the element number threshold is 7 for example. If the traversed text element is "light" (that is, the type of text element is a word), then it will be the first in the scanning cycle. In the second cycle, the number of scanning processes that have been executed is 0, and the number of scans is 6.
- the scanning process starting from “light” is executed to obtain "the convinced people of a certain country", and the combination of "glorious people of a certain country” "As a candidate entity word; in the second cycle of the scanning cycle, the number of scanning processes that have been executed is 1, the number of scans is 5, and the scanning process starting from "light” is executed according to the number of scans. People", and use the combination of "a elite countryman” as the candidate entity word, and so on, until the number of scans is equal to 0.
- the process of scanning according to the number of scans can be as follows: starting from the traversed text element, select the text element in the text to be recognized (the selected text element does not include the traversed text element), and use it as the scanning process. Until the total number of scanned text elements is equal to the number of scanned text elements. Wherein, when the number of scans on which a certain scan process is based is 0, since the text element obtained by the scan process is empty, the traversed text element itself can be used as the candidate entity word.
- the order of scanning processing can be from the first text element in the text to be recognized to the last text element (in the above example, this order is used as an example), or from the last text in the text to be recognized The order of the elements in turn to the first text element.
- the multiple candidate entity words obtained in step 202 may include duplicate candidate entity words. Therefore, all candidate entity words obtained can be deduplicated, that is, only one of the same multiple candidate entity words is retained to ensure After the deduplication process, the remaining multiple candidate entity words are different from each other, thereby saving subsequent computing resources.
- step 103 shown in FIG. 4A can be implemented through step 203 to step 204, which will be described in combination with each step.
- step 203 according to the sequence of the first text element in the to-be-recognized text to the last text element, the text representations of the multiple text elements in the candidate entity word are sequentially updated cyclically to obtain each text in the candidate entity word
- the text of the element represents the update result.
- the text representations of multiple text elements in the candidate entity word are sequentially updated cyclically to obtain each The update result of the text representation of the text element.
- the cyclic update process refers to the text representation of any text element in the candidate entity word, combined with the text representation of adjacent text elements, to update the text representation of any text element, and obtain the text representation of any text element
- the update result in this way, can improve the comprehensiveness and accuracy of the information contained in the obtained update result.
- adjacent can refer to the previous one, and can also include the previous one and the next one.
- the text representations of multiple text elements in the candidate entity words can be sequentially input into the RNN model, and the cyclic update processing is implemented based on the forward propagation process of the RNN model. Strong sequence memory performance and semantic representation ability, so it can improve the accuracy of the update results obtained.
- the RNN model can be a one-way RNN model or a two-way RNN model.
- the adjacent text element refers to the previous text element; in the two-way RNN model, the adjacent text element includes the previous text element And the next text element.
- the above-mentioned cyclic update processing of the text representations of multiple text elements in the candidate entity word can be implemented in this way to obtain the update result of the text representation of each text element in the candidate entity word:
- For the text representation of any text element in the candidate entity word perform the following operations: perform fusion processing on the first hidden state of the text representation of any text element and the text representation of the previous text element to obtain the text representation of any text element The first hidden layer state of any text element; the text representation of any text element and the second hidden layer state of the text representation of the following text element are fused to obtain the second hidden layer state of the text representation of any text element; for any text element The first hidden layer state and the second hidden layer state of the text representation of a text element are fused to obtain an update result of the text representation of any text element.
- cyclic update processing For ease of understanding, the process of performing cyclic update processing on the text representation of any text element in the candidate entity word is described.
- the text representation of any text element and the first hidden state (hidden state) of the text representation of the previous text element are fused to obtain the text representation of any text element
- the first hidden layer state of the fusion process can be a weighted summation, of course, it can also include other processing, such as adding a bias term to the result of the weighted summation, and then activate the result (by activating Function to realize activation processing, activation function such as hyperbolic tangent function, etc.).
- the text representation of any text element and the second hidden layer state of the text representation of the following text element are fused to obtain the second hidden layer state of the text representation of any text element.
- the first hidden layer state and the second hidden layer state of the text representation of any text element are fused to obtain the update result of the text representation of any text element.
- the text representation of the previous text element can be the first hidden state Set to zero; for the text representation of the last text element in the candidate entity word, since there is no text representation of the next text element, the second hidden layer state of the text representation of the next text element can be set to zero.
- step 204 the update result of the text representation of the last text element in the candidate entity word is used as the text representation of the candidate entity word.
- the update result of the text representation of the last text element in the candidate entity word it combines more information in the candidate entity word. Therefore, the update result is used as the text representation of the candidate entity word, so that the text of the candidate entity word It means that it can fully and effectively express the semantics of the candidate entity words.
- the embodiment of the present application constructs candidate entity words by traversing and scanning loops, which can be applied to the situation of multi-level nesting, and by performing loop update processing, the text representation of the obtained candidate entity words can be improved Accuracy and effectiveness.
- Figure 4C is an optional flowchart of the artificial intelligence-based named entity recognition method provided by an embodiment of the present application.
- Step 102 shown in Figure 4A can be updated to step 301.
- the following operations are performed according to multiple convolution windows of different lengths: perform a sliding operation of the convolution window in the text to be recognized, and the amplitude of each sliding operation is a text element;
- the covered text elements are combined into candidate entity words, and the text elements covered by the convolution window when the sliding stops are combined into candidate entity words; wherein the length of the convolution window is less than or equal to the element number threshold.
- the length of the convolution window Refers to the total number of text elements that can be covered by the convolution window.
- the length can be increased from the convolution window of length 1 until the convolution window whose length reaches the threshold of the number of elements is obtained. If the threshold of the number of elements is 7, set the length to 1 in turn 2.
- the convolution window of >, 7.
- a sliding operation of the convolution window is performed in the text to be recognized, and the amplitude of each sliding operation is one text element.
- the embodiment of the application does not limit the sequence of performing the sliding operation. For example, it may start from the first text element in the text to be recognized and slide to the last text element in the text to be recognized.
- the stop condition of the sliding operation ( Stop here refers to no more sliding operation) is the last text element in the text to be recognized by the convolution window; for example, it can also start from the last text element in the text to be recognized and move to the text to be recognized.
- the first text element of the sliding here, the stopping condition of the sliding operation is that the convolution window covers the first text element in the text to be recognized.
- the text to be recognized is "Glorious People's Liberation Army of a certain country", and the length of a certain convolution window is 3.
- the convolution window starting from the first text element in the text to be recognized, Slide the last text element, and before the first slide, use the “glorious” covered by the convolution window as the candidate entity word; before the second slide, the “glory” covered by the convolution window "Certain” as a candidate entity word, and so on.
- the convolution window covers "Liberation Army”
- step 103 shown in FIG. 4A can be updated to step 302.
- step 302 the text representation corresponding to the text element in the candidate entity word is convolved to obtain the text representation of the candidate entity word; where The size of the convolution kernel used for convolution processing is the same as the length of the convolution window used to construct the candidate entity words.
- convolution processing is performed on the text representations corresponding to all text elements in the candidate entity words to obtain the text representation of the candidate entity words, where the size of the convolution kernel used for the convolution processing (size) is consistent with the length of the convolution window used to construct the candidate entity words.
- the integration module 2433 for each convolution window, the text representations corresponding to all text elements in the candidate entity words constructed based on the convolution window are input into the CNN model, and the CNN model The output obtained by the forward propagation process is determined as the textual representation of the candidate entity word, where the size of the convolution kernel of the CNN model is consistent with the length of the convolution window.
- the embodiment of the present application constructs the candidate entity words by sliding the convolution window, realizes the construction of the candidate entity words from another angle, and convolves the text representations corresponding to the text elements in the candidate entity words. Processing to obtain the textual representation of the candidate entity word can improve the accuracy and effectiveness of the obtained textual representation of the candidate entity word.
- FIG. 4D is an optional flowchart of the artificial intelligence-based named entity recognition method provided in an embodiment of the present application.
- Step 104 shown in FIG. 4A can be implemented through steps 401 to 405 , Will be described in conjunction with each step.
- step 401 full connection processing is performed on the textual representation of the candidate entity word.
- the text representation of the candidate entity word is fully connected through the fully connected layer.
- the purpose of the full connection process is to extract and integrate the effective information in the text representation of the candidate entity word. Facilitate subsequent classification.
- step 402 the text representation of the candidate entity word after the fully connected process is mapped by the first classification function to obtain the probability corresponding to the non-named entity category and the probability of one-to-one correspondence with multiple named entity categories; where , The first classification function is used to classify the candidate entity words twice.
- the first classification function is used to map the text representations of the candidate entity words after the fully-connected process to obtain the NAND
- the second classification in the embodiment of the present application refers to the case where there are no named entities of multiple types, and it is judged whether the candidate entity word belongs to the non-named entity category or the named entity category.
- step 403 the category corresponding to the probability with the largest numerical value is determined as the category to which the candidate entity word belongs.
- the category corresponding to the probability with the largest numerical value is determined as the category to which the candidate entity word belongs, that is, the category to which the candidate entity word belongs can only be a non-named entity category or a named entity category.
- a second classification function is used to perform a mapping process on the text representation of the candidate entity word after the fully connected process, to obtain the probability corresponding to the non-named entity category and the probability of one-to-one correspondence with multiple named entity categories.
- the second classification function is used to map the text representations of the candidate entity words after the fully connected process, and the The probability corresponding to the entity category and the probability corresponding to multiple named entity categories one-to-one, wherein the second classification function may be a Sigmoid classification function.
- the multi-category in the embodiment of the present application refers to the case where there are multiple categories of named entities, and it is judged whether the candidate entity word belongs to a non-named entity category or at least one named entity category.
- step 405 the category corresponding to the probability exceeding the probability threshold is determined as the category to which the candidate entity word belongs.
- a probability threshold is set, and the category corresponding to the probability exceeding the probability threshold is determined as the category to which the candidate entity word belongs.
- the candidate entity word may belong to one or more categories.
- the embodiment of the present application uses different classification functions for classification processing for the two-classification and multi-classification situations, which improves the flexibility of named entity recognition.
- the embodiment of the present application provides an optional schematic diagram of the architecture of using the recurrent neural network model for named entity recognition as shown in FIG.
- This module generally refers to the NLP model structure with text representation capabilities and other extensions based on the representation structure, which can be set according to actual application scenarios. For example, the module can apply the BERT model and its improvements. Input the text to be recognized into the module to obtain the text representation of the text element in the text to be recognized, where the type of the text element is a word or word, and the case where the type of the text element is a word is illustrated in FIG. 5 as an example.
- the text to be recognized is "Glorious People's Liberation Army of a certain country”.
- the text representation of each word in the text to be recognized is obtained in the form of a vector.
- the box is an example.
- the RNN model in the embodiments of the present application generally refers to the RNN model and its variants, such as the Long Short-Term Memory (LSTM) model and the Gated Recurrent Unit (GRU) model.
- LSTM Long Short-Term Memory
- GRU Gated Recurrent Unit
- a two-way RNN model is taken as an example.
- Each circle in the model represents a cell at a certain moment, where the input of the cell is a text representation of a text element.
- it is used to pass the first hidden layer state of the input text to the next cell, and it is also used to pass the second hidden layer state of the input text to the previous cell.
- the process of constructing candidate entity words can be divided into two layers of processing, the first layer is traversal processing, and the second layer is scanning loop.
- the traversal processing is performed in the order from the first text element to the last text element in the text to be recognized.
- the purpose of the traversal processing is to completely cover the beginning of all possible candidate entity words in the text to be recognized .
- the text representation corresponding to the traversed text element is used as the input vector at the first moment (step) in the RNN model. It is worth noting that for the different text elements traversed, the input to the RNN model is performed independently of each other, that is, the text representation corresponding to the text element traversed before will not be used as the input of the next RNN model. The text representation corresponding to a traversed text element will reuse the RNN model, that is, as the input vector at the first moment of the new RNN model.
- the traversed text element is " ⁇ "
- the text corresponding to " ⁇ ” is used as the input vector at the first moment in the RNN model
- the traversed text element is " ⁇ "
- the corresponding text representation is used as the input vector at the first moment in another RNN model.
- the scanning cycle refers to the beginning of the candidate entity word with the beginning of the word, and according to the gradually increasing number of scans (the number of scans is initially 0), the cycle scans each candidate entity word starting with the beginning of the word, until the number of scans reaches the scan The number threshold, where the scan number threshold is the result of the element number threshold minus one. For example, if the traversed text element (the beginning of the word) is "some", the scanning cycle is started, that is, the word "some” is used as the beginning of the candidate entity word, and the number of scans is gradually expanded.
- While constructing the candidate entity words input the text representations corresponding to the scanned words one by one into the RNN model one by one, and use the output of the RNN time corresponding to the currently scanned word (corresponding to the update result above) , As the textual representation of the candidate entity words from the "first word” to the current scanned word. For example, for the candidate entity word "a person from a country”, the text representations corresponding to "some", “country” and “person” are input into the RNN model one by one in turn, and the output at the time corresponding to "people” is outputted. Determined as the textual representation of the candidate entity word "a person from a certain country”.
- the sequence memory performance and semantic representation ability of the RNN model can be better utilized, so that the obtained text representation of the candidate entity word can accurately and effectively represent the semantics of the candidate entity word.
- Figure 5 shows multiple candidate entity words constructed after traversal processing and scanning loops. For each candidate entity word, the forward propagation processing of the RNN model can be used to obtain the textual representation of the candidate entity word.
- the classification layer is used to classify the text representations of multiple candidate entity words output by the RNN model.
- the embodiment of the present application provides two classification methods, using two activation (classification) functions, Softmax and Sigmoid, respectively.
- the Softmax activation function can be used, and the text representation of each candidate entity word obtained is used to classify using a structure such as a fully connected layer and the Softmax activation function. , And finally get a category to which the candidate entity word is most likely to belong.
- the Sigmoid activation function can be used, and the fully connected layer and other structures are combined with the Sigmoid activation function for classification. Finally, the candidate entity is obtained by setting the probability threshold. The category to which the word may belong. It is worth noting that in the classification process, non-named entity categories can be treated equally with other named entity categories.
- the category to which the candidate entity word "some” belongs is a non-named entity category
- the category to which the candidate entity word "some country” belongs is a country name category
- the candidate entity word The category to which "people from a certain country” belongs is a group category
- the category to which the candidate entity word "People's Liberation Army of a certain country” belongs is an army category.
- the obtained named entity can be applied to various application scenarios in the NLP field. For example, in the scene of new word discovery, the obtained named entity can be used Add to the knowledge graph to realize automatic mining of new named entities.
- the embodiment of the present application also provides an optional schematic diagram of the architecture of using the convolutional neural network model for named entity recognition as shown in FIG. 6, which will be explained in order from bottom to top.
- the content is similar to the corresponding content in FIG. 5.
- the CNN model in this part generally refers to the CNN model and its variants.
- Figure 6 takes the original CNN model as an example.
- a CNN model with multiple convolution kernel sizes is used to perform one-dimensional CNN processing on the text representation of the text elements in the candidate entity words.
- the sizes are from 1 to 7.
- the number of convolution kernels can be set according to specific actual application scenarios. Convolutions of different sizes The number of cores can be set to be the same.
- set a convolution window with the same length as the size of the convolution kernel and perform a sliding operation on the convolution window according to the sequence of the first text element in the text to be recognized to the last text element.
- the text elements covered by the convolution window before each sliding are combined as candidate entity words, and the text elements covered by the convolution window when the sliding stops are combined as candidate entity words.
- the text representations corresponding to the text elements in the candidate entity words are uniformly input to the CNN model with the corresponding size of the convolution kernel, and the convolution output result of the CNN model is the text representation of the candidate entity words.
- candidate entity words constructed by "a certain”, “a certain country”, and “a certain country person” are shown.
- the Softmax/Sigmoid classification layer in Figure 6 is similar to the corresponding content in Figure 5, and different activation functions can also be used for different classification tasks.
- the artificial intelligence-based named entity recognition method provided by the embodiments of this application eliminates the need to manually formulate complex coding and decoding rules, which improves the efficiency and flexibility of named entity recognition.
- the embodiments of the present application also have certain advantages.
- the model structure of the embodiment of the present application is simpler, and at the same time, it can more effectively enhance the semantic expression ability of candidate entity words.
- Improve the performance indicators of the model Experiments have verified that it is tested under the same text representation structure based on language model embedding (ELMo, Embedding from Language Models). The indicators of the two schemes are compared as follows:
- the F1 score of the MGNER model scheme is 79.5%, and the F1 score of the method provided in the embodiment of this application is 83.7%; under the ACE2005 public data set, the F1 score of the MGNER model scheme is 78.2%.
- the F1 score of the method provided in the application example is 82.4%, where the F1 score is the harmonic average of the precision rate and the recall rate.
- machine reading comprehension can also be used to realize named entity recognition, but this solution requires a model to be run independently for each type of named entity. It takes a long time.
- all possible candidate entity words and their categories in the text to be recognized can be generated in one run, which is more efficient.
- the ACL-NNE public data set with 115 named entity types is used for testing, and when the MRC framework solution is also based on the BERT model, compared with the MRC framework solution, the method provided in the embodiments of this application can save about 90% %time.
- the results of named entity recognition can be applied to various application scenarios in the field of NLP, such as summary determination, object recommendation, text classification, and Q&A system scenarios, as well as information extraction, grammatical analysis and machine translation scenarios, here is the Q&A system scenario Give a detailed description.
- the embodiment of the present application provides a schematic diagram of the question and answer process as shown in FIG. 7.
- the terminal device 400-1 and the terminal device 400-2 are held by different users.
- the terminal device 400- 1 is named the first terminal device
- the terminal device 400-2 is named the second terminal device.
- the question and answer process will be described in conjunction with the steps shown in FIG. 7.
- step 501 the first terminal device sends the text to be recognized to the server.
- the user of the first terminal device sends the text to be recognized to the server through the first terminal device.
- the source of the text to be recognized is not limited here.
- the text to be recognized can be the entry text of a certain person or the description of a certain product. Text etc.
- step 502 the server constructs a candidate entity word according to the text elements in the text to be recognized, and determines the category to which the candidate entity word belongs.
- the server can construct all possible candidate entity words in an exhaustive manner, and determine the category to which each candidate entity word belongs. Refer to the above step 101 to step 104.
- the server determines the candidate entity words belonging to the named entity category as keywords, and performs syntactic analysis on the text to be recognized to obtain subject keywords, relation words, and object keywords.
- the server determines all candidate entity words belonging to any named entity category as keywords, and further performs syntactic analysis on the text to be recognized to obtain subject keywords, relative words, and object keywords that have dependencies in the text to be recognized , Among them, the relative words are used to express the relationship between the subject keywords and the object keywords. For example, if the text to be recognized is "Zhang San borrows money from Li Si", through syntactic analysis, the subject keyword is "Zhang San", the relative word is "borrow money”, and the object keyword is "Li Si”.
- step 504 the server constructs a triplet according to the subject keywords, relative words, and object keywords, and adds the triples to the knowledge graph.
- the subject-predicate-object triples are constructed according to the subject keywords, relative words, and object keywords.
- the subject-predicate-object triplet is "Zhang San-Borrow Money-Li Si".
- the server adds the constructed subject-predicate-object triples to the knowledge graph.
- "Zhang San” and “Li Si” can also be added to the knowledge graph as named entities.
- step 505 the second terminal device sends an object query request including subject keywords and related words to the server.
- the second terminal device sends an object query request of "Zhang San borrowed money from?” to the server.
- step 506 the server determines the semantics of the object query request through the knowledge graph, and performs a query in the knowledge graph according to the semantics to obtain the query result.
- the server can match the object query request with the named entity in the knowledge graph, for example, match "Zhang San borrowed money from?” with the named entity in the knowledge graph to obtain the name matching the knowledge graph in the object query request The entity is "Zhang San”.
- the server can also apply the method of step 101 to step 104 to obtain the category to which the candidate entity word in the object query request belongs, and match the candidate entity word belonging to the named entity category with the knowledge graph.
- the server further performs syntactic analysis and processing according to the named entity matching the knowledge graph in the object query request to obtain the semantics of the query request, such as "Zhang San-borrow money-?".
- the server performs a query in the knowledge graph according to the semantics of the object query request, obtains the corresponding query result, and sends the query result to the second terminal device to complete the question and answer, for example, the object keyword "Li Si" is sent as the query result to The second terminal device.
- the embodiment of the present application expands the knowledge graph based on the result of named entity recognition, which improves the accuracy of knowledge in the knowledge graph, and also improves the accuracy of the question answering system based on the knowledge graph, so that users can participate in question and answer Get a good user experience.
- the artificial intelligence-based named entity recognition device 243 may include: a vector conversion module 2431, configured to perform vector conversion processing on text elements in the text to be recognized, to obtain a text representation of the text elements; a construction module 2432, configured to vary according to the total number of text to be recognized The text elements that exceed the threshold of the number of elements are used to construct candidate entity words; the integration module 2433 is configured to integrate the text representations corresponding to the text elements in the candidate entity words to obtain the text representations of the candidate entity words; the classification module 2434 is configured to The text representation of the entity word is classified to determine the category to which the candidate entity word belongs among the non-named entity category and multiple named entity categories.
- the construction module 2432 is configured to: perform traversal processing on the text elements in the text to be recognized; perform a scanning cycle including multiple scanning processing for the traversed text elements: according to the scans that have been performed in the scanning cycle The number of processing, determine the number of scans to be increased or decreased simultaneously, execute the scan processing from the traversed text element according to the scan number, and combine the traversed text element and the text element obtained by the scanning process into candidate entity words, The number of scans until the initial zero is increased to the threshold of the scan number, or the number of scans that is initially the threshold of the scan number is reduced to zero; wherein the threshold of the number of scans is the result obtained by subtracting one from the threshold of the number of elements.
- the integration module 2433 is configured to: according to the sequence of the first text element in the to-be-recognized text to the last text element, the text representations of the multiple text elements in the candidate entity word are sequentially updated cyclically, Obtain the update result of the text representation of each text element in the candidate entity word; take the update result of the text representation of the last text element in the candidate entity word as the text representation of the candidate entity word.
- the integration module 2433 is configured to perform the following operations for the text representation of any text element in the candidate entity word: the text representation of any text element and the first hidden text representation of the previous text element
- the layer state is fused to obtain the first hidden layer state of the text representation of any text element; the text representation of any text element and the second hidden layer state of the text representation of the following text element are fused to obtain any one
- the second hidden layer state of the text representation of the text element; the first hidden layer state and the second hidden layer state of the text representation of any text element are fused to obtain the update result of the text representation of any text element.
- the construction module 2432 is configured to select the text elements in the text to be recognized starting from the traversed text elements as the text elements obtained by the scanning process, until the total number of the text elements obtained by the scanning process is equal to the scanning process. quantity.
- the construction module 2432 is configured to: perform the following operations according to multiple convolution windows of different lengths: perform a sliding operation of the convolution window in the text to be recognized, and the amplitude of each sliding operation is one text element ; Combine the text elements covered by the convolution window before each sliding as candidate entity words, and combine the text elements covered by the convolution window when sliding is stopped as candidate entity words; where the length of the convolution window is less than or equal to the element The number threshold.
- the integration module 2433 is configured to: perform convolution processing on the text representation corresponding to the text element in the candidate entity word to obtain the text representation of the candidate entity word; wherein, the size of the convolution kernel used for the convolution processing Consistent with the length of the convolution window used to construct candidate entity words.
- the classification module 2434 is configured to: perform full connection processing on the text representation of the candidate entity word; perform mapping processing on the text representation of the candidate entity word after the full connection processing through the first classification function, to obtain the ND
- the probability corresponding to the entity category and the probability corresponding to multiple named entity categories one-to-one; the category corresponding to the probability with the largest numerical value is determined as the category to which the candidate entity word belongs; where the first classification function is used to perform the classification of the candidate entity word Two classification.
- the classification module 2434 is configured to: perform full connection processing on the text representation of the candidate entity word; perform mapping processing on the text representation of the candidate entity word after the full connection processing through the second classification function, to obtain the NAND and non-named
- the probability corresponding to the entity category and the probability of one-to-one correspondence with multiple named entity categories; the category corresponding to the probability exceeding the probability threshold is determined as the category to which the candidate entity word belongs; where the second classification function is used to compare the candidate entity word Perform multiple classifications.
- the named entity recognition device 243 based on artificial intelligence further includes: a segmentation module configured to perform segmentation processing on the text to be recognized to obtain a plurality of sentences; , And the candidate entity words whose appearance frequency meets the frequency condition are determined as summary keywords; the score determination module is configured to determine the score of the sentence according to the number of summary keywords included in the sentence; the summary determination module is configured to determine the score that meets the scoring condition The sentence is determined as the text summary of the text to be recognized.
- the named entity recognition device 243 based on artificial intelligence further includes: a first keyword determination module configured to identify candidate entity words belonging to any named entity category when the text to be recognized is used to represent the object to be recommended , Determine as the keyword of the object to be recommended; the user portrait acquisition module, configured to acquire the keyword of the user portrait, and determine the degree of keyword overlap between the keyword of the user portrait and the keyword of the object to be recommended; the recommendation module, configured to When the keyword coincidence degree exceeds the first coincidence degree threshold, the operation of recommending the object to be recommended is performed.
- a first keyword determination module configured to identify candidate entity words belonging to any named entity category when the text to be recognized is used to represent the object to be recommended , Determine as the keyword of the object to be recommended
- the user portrait acquisition module configured to acquire the keyword of the user portrait, and determine the degree of keyword overlap between the keyword of the user portrait and the keyword of the object to be recommended
- the recommendation module configured to When the keyword coincidence degree exceeds the first coincidence degree threshold, the operation of recommending the object to be recommended is performed.
- the user portrait acquisition module is configured to: determine the intersection between the user portrait keywords and the keywords of the object to be recommended, and determine the first number of keywords included in the intersection; determine the user portrait keywords and the keywords to be recommended. The union between the keywords of the recommended object, and the second number of keywords included in the union is determined; the ratio between the first number and the second number is determined as the keyword of the user portrait and the keyword of the object to be recommended The degree of overlap between keywords.
- the named entity recognition device 243 based on artificial intelligence further includes: a second keyword determination module configured to determine candidate entity words belonging to any named entity category as keywords; and a coincidence degree calculation module configured to Determine the keyword overlap between the first text to be recognized and the second text to be recognized; the classification module is configured to compare the first text to be recognized with the second text to be recognized when the keyword overlap exceeds the second overlap threshold. The text is divided into the same text category.
- the named entity recognition device 243 based on artificial intelligence further includes: a third keyword determination module configured to determine candidate entity words belonging to any named entity category as keywords; a syntax analysis module configured to treat Recognize the text for syntactic analysis to obtain the subject keywords, relative words and object keywords in the text to be recognized; among them, the relative words are used to express the relationship between the subject keywords and the object keywords; add a module and configure it according to the subject Keywords, relative words, and object keywords construct triples, and add the triples to the knowledge graph; among them, the knowledge graph is used to respond to object query requests including subject keywords and relative words.
- a third keyword determination module configured to determine candidate entity words belonging to any named entity category as keywords
- a syntax analysis module configured to treat Recognize the text for syntactic analysis to obtain the subject keywords, relative words and object keywords in the text to be recognized; among them, the relative words are used to express the relationship between the subject keywords and the object keywords
- the named entity recognition device 243 based on artificial intelligence further includes: an entity adding module configured to add candidate entity words belonging to any named entity category to the knowledge graph; wherein the knowledge graph is used to respond to the The category query request of the candidate entity word of the entity category.
- the named entity recognition device 243 based on artificial intelligence further includes: an element recognition module configured to perform any one of the following operations to obtain text elements in the text to be recognized: each word in the text to be recognized They are all used as text elements; the text to be recognized is subjected to word segmentation processing, and the words obtained by word segmentation processing are used as text elements.
- an element recognition module configured to perform any one of the following operations to obtain text elements in the text to be recognized: each word in the text to be recognized They are all used as text elements; the text to be recognized is subjected to word segmentation processing, and the words obtained by word segmentation processing are used as text elements.
- the embodiment of the present application provides a computer-readable storage medium storing executable instructions, and the executable instructions are stored therein.
- the processor will cause the processor to execute the method provided in the embodiments of the present application, for example, , As shown in FIG. 4A, FIG. 4B, FIG. 4C, or FIG. 4D, a named entity recognition method based on artificial intelligence.
- computers include various computing devices including terminal devices and servers.
- the computer-readable storage medium may be FRAM, ROM, PROM, EPROM, EEPROM, flash memory, magnetic surface memory, optical disk, or CD-ROM, etc.; it may also include one or any combination of the foregoing memories.
- FRAM fast access memory
- ROM read-only memory
- PROM PROM
- EPROM erasable programmable read-only memory
- EEPROM electrically erasable programmable read-only memory
- flash memory magnetic surface memory
- optical disk optical disk
- CD-ROM compact flash memory
- the executable instructions may be in the form of programs, software, software modules, scripts or codes, written in any form of programming language (including compiled or interpreted languages, or declarative or procedural languages), and their It can be deployed in any form, including being deployed as an independent program or as a module, component, subroutine or other unit suitable for use in a computing environment.
- executable instructions may but do not necessarily correspond to files in the file system, and may be stored as part of files that store other programs or data, for example, in a HyperText Markup Language (HTML, HyperText Markup Language) document
- HTML HyperText Markup Language
- One or more scripts in are stored in a single file dedicated to the program in question, or in multiple coordinated files (for example, a file storing one or more modules, subroutines, or code parts).
- executable instructions can be deployed to be executed on one computing device, or on multiple computing devices located in one location, or on multiple computing devices that are distributed in multiple locations and interconnected by a communication network Executed on.
- the embodiment of the application provides a refined and easy-to-use method for constructing candidate entity words, which integrates sequence information in the text to be recognized one by one and in different spans in a specific order, and can fully enumerate the words in the text to be recognized.
- candidate entity words The model of the embodiment of the present application has a simple structure and strong flexibility, which is convenient for further improvement according to the needs in actual application scenarios, and is also easy to be transplanted to more deep learning models.
- the embodiments of this application conform to the application characteristics of the RNN model and the CNN model in the NLP field.
- the relevant structure of the RNN model or the CNN model is used to integrate the text representations of the text elements in the text to be recognized to obtain candidates
- the text representation of entity words can more simply and effectively enhance the semantic expression ability of candidate entity words, improve the performance indicators of the model, and give consideration to simplicity and effectiveness.
- the embodiments of the present application use corresponding classification functions for classification processing, which improves the applicability to different application scenarios.
- the importance of each sentence can be determined according to the candidate entity words belonging to the named entity category, so as to filter out the text abstract and improve the selection of abstract accuracy.
- the embodiment of this application matches the keywords in the text to be recognized with the keywords of the user portrait, so as to recommend objects that meet the user's preferences as much as possible, which improves the user experience and the recommended The conversion rate of the object.
- the embodiment of the application compares the keywords between the two texts, and judges whether to classify the two texts into one category according to the obtained keyword overlap, which improves the accuracy of text classification.
- candidate entity words belonging to the named entity category can be added to the knowledge graph to improve the accuracy of new word discovery.
- relationship between named entities appearing in the text to be recognized can also be added to the knowledge graph, so that the expanded knowledge graph can be better applied to scenarios such as question answering.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Databases & Information Systems (AREA)
- Probability & Statistics with Applications (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Animal Behavior & Ethology (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
In | O |
the | B-ORG |
US | I-ORG|U-GPE |
Federal | I-ORG |
District | I-ORG|U-GPE |
Court | I-ORG |
of | I-ORG |
New | I-ORG|B-GPE |
Mexico | I-ORG|L-GPE |
. | O |
Claims (19)
- 一种基于人工智能的命名实体识别方法,由电子设备执行,包括:对待识别文本中的文本元素进行向量转换处理,得到所述文本元素的文本表示;根据所述待识别文本中总数量不超过元素数量阈值的文本元素,构建候选实体词;对所述候选实体词中文本元素对应的文本表示进行整合处理,得到所述候选实体词的文本表示;对所述候选实体词的文本表示进行分类处理,以在非命名实体类别和多个命名实体类别中,确定所述候选实体词所属的类别。
- 根据权利要求1所述的命名实体识别方法,其中,所述根据所述待识别文本中总数量不超过元素数量阈值的文本元素,构建候选实体词,包括:对所述待识别文本中的文本元素进行遍历处理;针对遍历到的文本元素,执行包括多次扫描处理的扫描循环:根据在所述扫描循环中已经执行的扫描处理的次数,确定同步增大或减小的扫描数量,根据所述扫描数量执行从所述遍历到的文本元素开始的扫描处理,并将所述遍历到的文本元素和扫描处理得到的文本元素组合为候选实体词,直至初始为零的所述扫描数量增大至扫描数量阈值、或者初始为所述扫描数量阈值的所述扫描数量减小至零;其中,所述扫描数量阈值为所述元素数量阈值减一后得到的结果。
- 根据权利要求2所述的命名实体识别方法,其中,所述对所述候选实体词中文本元素对应的文本表示进行整合处理,得到所述候选实体词的文本表示,包括:根据所述待识别文本中第一个文本元素依次到最后一个文本元素的顺序,对所述候选实体词中多个文本元素的文本表示依次进行循环更新处理,得到所述候选实体词中每个文本元素的文本表示的更新结果;将所述候选实体词中最后一个文本元素的文本表示的更新结果,作为所述候选实体词的文本表示。
- 根据权利要求3所述的命名实体识别方法,其中,所述对所述候选实体词中多个文本元素的文本表示依次进行循环更新处理,得到所述候选实体词中每个文本元素的文本表示的更新结果,包括:针对所述候选实体词中任意一个文本元素的文本表示,执行以下操作:对所述任意一个文本元素的文本表示与前一个文本元素的文本表示的第一隐层状态进行融合处理,得到所述任意一个文本元素的文本表示的第一隐层状态;对所述任意一个文本元素的文本表示与后一个文本元素的文本表示的第二隐层状态进行融合处理,得到所述任意一个文本元素的文本表示的第二隐层状态;对所述任意一个文本元素的文本表示的第一隐层状态及第二隐层状态进行融合处理,得到所述任意一个文本元素的文本表示的更新结果。
- 根据权利要求2所述的命名实体识别方法,其中,所述根据所述扫描数量执行从所述遍历到的文本元素开始的扫描处理,包括:从所述遍历到的文本元素开始选取所述待识别文本中的文本元素,以作为扫描处理得到的文本元素,直至扫描处理得到的文本元素的总数量等于所述扫描数量。
- 根据权利要求1所述的命名实体识别方法,其中,所述根据所述待识别文本中总数量不超过元素数量阈值的文本元素,构建候选实体词,包括:根据多个不同长度的卷积窗口执行以下操作:在所述待识别文本中执行所述卷积窗口的滑动操作,且每次滑动操作的幅度为一个文本元素;将每次滑动前所述卷积窗口所覆盖的文本元素组合为候选实体词,并将停止滑动时所述卷积窗口所覆盖的文本元素组合为候选实体词;其中,所述卷积窗口的长度小于或等于所述元素数量阈值。
- 根据权利要求6所述的命名实体识别方法,其中,所述对所述候选实体词中文本元素对应的文本表示进行整合处理,得到所述候选实体词的文本表示,包括:对所述候选实体词中文本元素对应的文本表示进行卷积处理,得到所述候选实体词的文本表示;其中,用于进行卷积处理的卷积核尺寸与用于构建所述候选实体词的所述卷积窗口的长度一致。
- 根据权利要求1所述的命名实体识别方法,其中,所述对所述候选实体词的文本表示进行分类处理,以在非命名实体类别和多个命名实体类别中,确定所述候选实体词所属的类别,包括:对所述候选实体词的文本表示进行全连接处理;通过第一分类函数对全连接处理后的所述候选实体词的文本表示进行映射处理,得到与所述非命名实体类别对应的概率、以及与所述多个命名实体类别一一对应的概率;将数值最大的概率对应的类别,确定为所述候选实体词所属的类别;其中,所述第一分类函数用于对所述候选实体词进行二分类。
- 根据权利要求1所述的命名实体识别方法,其中,所述对所述候选实体词的文本表示进行分类处理,以在非命名实体类别和多个命名实体类别中,确定所述候选实体词所属的类别,包括:对所述候选实体词的文本表示进行全连接处理;通过第二分类函数对全连接处理后的所述候选实体词的文本表示进行映射处理,得到与所述非命名实体类别对应的概率、以及与所述多个命名实体类别一一对应的概率;将超过概率阈值的概率对应的类别,确定为所述候选实体词所属的类别;其中,所述第二分类函数用于对所述候选实体词进行多分类。
- 根据权利要求1至9任一项所述的命名实体识别方法,其中,还包括:对所述待识别文本进行分割处理得到多个语句;将属于任意一个所述命名实体类别的、且出现频率满足频率条件的候选实体词确定为摘要关键词;根据所述语句包括的摘要关键词的数量,确定所述语句的评分;将评分满足评分条件的语句,确定为所述待识别文本的文本摘要。
- 根据权利要求1至9任一项所述的命名实体识别方法,其中,还包括:当所述待识别文本用于表示待推荐对象时,将属于任意一个所述命名实体类别的候选实体词,确定为所述待推荐对象的关键词;获取用户画像关键词,并确定所述用户画像关键词与所述待推荐对象的关键词之间的关键词重合度;当所述关键词重合度超过第一重合度阈值时,执行推荐所述待推荐对象的操作。
- 根据权利要求11所述的命名实体识别方法,其中,所述确定所述用户画像关键词与所述待推荐对象的关键词之间的关键词重合度,包括:确定所述用户画像关键词与所述待推荐对象的关键词之间的交集,并确定所述交集包括的关键词的第一数量;确定所述用户画像关键词与所述待推荐对象的关键词之间的并集,并确定所述并集包括的关键词的第二数量;将所述第一数量与所述第二数量之间的比值,确定为所述用户画像关键词与所述待推荐对象的关键词之间的关键词重合度。
- 根据权利要求1至9任一项所述的命名实体识别方法,其中,还包括:将属于任意一个所述命名实体类别的候选实体词确定为关键词;确定第一待识别文本与第二待识别文本之间的关键词重合度;当所述关键词重合度超过第二重合度阈值时,将所述第一待识别文本与所述第二待识别文本划分为同一个文本类。
- 根据权利要求1至9任一项所述的命名实体识别方法,其中,还包括:将属于任意一个所述命名实体类别的候选实体词确定为关键词;对所述待识别文本进行句法分析处理,得到所述待识别文本中的主语关键词、关系词及宾语关键词;其中,所述关系词用于表示所述主语关键词与所述宾语关键词之间的关系;根据所述主语关键词、所述关系词及所述宾语关键词构建三元组,并将所述三元组添加至知识图谱;其中,所述知识图谱用于响应包括所述主语关键词及所述关系词的宾语查询请求。
- 根据权利要求1至9任一项所述的命名实体识别方法,其中,还包括:将属于任意一个所述命名实体类别的候选实体词添加至知识图谱;其中,所述知识图谱用于响应针对属于任意一个所述命名实体类别的候选实体词的类别查询请求。
- 根据权利要求1至9任一项所述的命名实体识别方法,其中,还包括:执行以下任意一种操作,以得到所述待识别文本中的文本元素:将所述待识别文本中的每个字均作为文本元素;对所述待识别文本进行分词处理,将分词处理得到的词作为文本元素。
- 一种基于人工智能的命名实体识别装置,包括:向量转换模块,配置为对待识别文本中的文本元素进行向量转换处理,得到所述文本元素的文本表示;构建模块,配置为根据所述待识别文本中总数量不超过元素数量阈值的文本元素,构建候选实体词;整合模块,配置为对所述候选实体词中文本元素对应的文本表示进行整合处理,得到所述候选实体词的文本表示;分类模块,配置为对所述候选实体词的文本表示进行分类处理,以在非命名实体类别和多个命名实体类别中,确定所述候选实体词所属的类别。
- 一种电子设备,包括:存储器,用于存储可执行指令;处理器,用于执行所述存储器中存储的可执行指令时,实现权利要求1至16任一项所述的基于人工智能的命名实体识别方法。
- 一种计算机可读存储介质,存储有可执行指令,用于引起处理器执行时,实现权利要求1至16任一项所述的基于人工智能的命名实体识别方法。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/685,283 US20220188521A1 (en) | 2020-02-28 | 2022-03-02 | Artificial intelligence-based named entity recognition method and apparatus, and electronic device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010127101.0A CN111353310B (zh) | 2020-02-28 | 2020-02-28 | 基于人工智能的命名实体识别方法、装置及电子设备 |
CN202010127101.0 | 2020-02-28 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/685,283 Continuation US20220188521A1 (en) | 2020-02-28 | 2022-03-02 | Artificial intelligence-based named entity recognition method and apparatus, and electronic device |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021169400A1 true WO2021169400A1 (zh) | 2021-09-02 |
Family
ID=71194185
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/127737 WO2021169400A1 (zh) | 2020-02-28 | 2020-11-10 | 基于人工智能的命名实体识别方法、装置及电子设备 |
Country Status (3)
Country | Link |
---|---|
US (1) | US20220188521A1 (zh) |
CN (1) | CN111353310B (zh) |
WO (1) | WO2021169400A1 (zh) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114036933A (zh) * | 2022-01-10 | 2022-02-11 | 湖南工商大学 | 基于法律文书的信息抽取方法 |
CN114048293A (zh) * | 2022-01-11 | 2022-02-15 | 广东拓思软件科学园有限公司 | 一种缺陷报告融合方法、装置、电子设备及存储介质 |
CN114841163A (zh) * | 2022-03-31 | 2022-08-02 | 阿里巴巴(中国)有限公司 | 实体识别方法、装置、设备及存储介质 |
CN116822632A (zh) * | 2023-08-28 | 2023-09-29 | 腾讯科技(深圳)有限公司 | 一种文本数据的推理方法、装置、存储介质和电子设备 |
Families Citing this family (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111353310B (zh) * | 2020-02-28 | 2023-08-11 | 腾讯科技(深圳)有限公司 | 基于人工智能的命名实体识别方法、装置及电子设备 |
CN111930939A (zh) * | 2020-07-08 | 2020-11-13 | 泰康保险集团股份有限公司 | 一种文本检测的方法及装置 |
CN111914101B (zh) * | 2020-08-17 | 2023-10-20 | 南方电网数字电网研究院有限公司 | 档案关联关系的异常识别方法、装置和计算机设备 |
RU2760637C1 (ru) * | 2020-08-31 | 2021-11-29 | Публичное Акционерное Общество "Сбербанк России" (Пао Сбербанк) | Способ и система извлечения именованных сущностей |
CN112560459B (zh) * | 2020-12-04 | 2023-10-20 | 北京百度网讯科技有限公司 | 用于模型训练的样本筛选方法、装置、设备及存储介质 |
CN112528600B (zh) * | 2020-12-15 | 2024-05-07 | 北京百度网讯科技有限公司 | 文本数据处理方法、相关装置及计算机程序产品 |
CN114722817A (zh) * | 2020-12-22 | 2022-07-08 | 北京金山数字娱乐科技有限公司 | 事件处理方法及装置 |
US11675978B2 (en) * | 2021-01-06 | 2023-06-13 | International Business Machines Corporation | Entity recognition based on multi-task learning and self-consistent verification |
CN112765994A (zh) * | 2021-01-26 | 2021-05-07 | 武汉大学 | 一种基于深度学习的信息要素联合抽取方法及系统 |
CN113326701A (zh) * | 2021-06-17 | 2021-08-31 | 广州华多网络科技有限公司 | 嵌套实体识别方法、装置、计算机设备及存储介质 |
CN113378572B (zh) * | 2021-06-22 | 2023-11-10 | 云知声智能科技股份有限公司 | 一种命名实体识别方法、装置、电子设备和存储介质 |
CN113516129B (zh) * | 2021-07-15 | 2024-07-16 | 广州云从鼎望科技有限公司 | 名片的识别方法及装置、计算机可读存储介质、控制装置 |
CN113642331B (zh) * | 2021-08-10 | 2022-05-03 | 东方财富信息股份有限公司 | 金融命名实体识别方法及系统、存储介质及终端 |
CN113377930B (zh) * | 2021-08-13 | 2021-11-30 | 中国科学院自动化研究所 | 面向中文短文本的实体识别与实体链接方法 |
CN113836874A (zh) * | 2021-09-16 | 2021-12-24 | 北京小米移动软件有限公司 | 文本纠错方法及装置 |
CN113887232B (zh) * | 2021-12-07 | 2022-02-22 | 北京云迹科技有限公司 | 一种对话信息的命名实体识别方法、装置和电子设备 |
US20230359820A1 (en) * | 2022-05-03 | 2023-11-09 | Sap Se | Natural language processing to extract skills characterization |
CN115221882B (zh) * | 2022-07-28 | 2023-06-20 | 平安科技(深圳)有限公司 | 命名实体识别方法、装置、设备及介质 |
CN116092493B (zh) * | 2023-04-07 | 2023-08-25 | 广州小鹏汽车科技有限公司 | 语音交互方法、服务器和计算机可读存储介质 |
KR102674954B1 (ko) * | 2023-10-20 | 2024-06-14 | 주식회사 마인즈앤컴퍼니 | 거대 언어 모델과 딥러닝을 이용한 검색 서비스를 제공하는 방법 및 장치 |
CN117609403A (zh) * | 2023-10-30 | 2024-02-27 | 合肥工业大学 | 一种应用于服务热线智能派单的地名识别方法及系统 |
CN118114675B (zh) * | 2024-04-29 | 2024-07-26 | 支付宝(杭州)信息技术有限公司 | 基于大语言模型的医疗命名实体识别方法和装置 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106547733A (zh) * | 2016-10-19 | 2017-03-29 | 中国国防科技信息中心 | 一种面向特定文本的命名实体识别方法 |
CN109101481A (zh) * | 2018-06-25 | 2018-12-28 | 北京奇艺世纪科技有限公司 | 一种命名实体识别方法、装置及电子设备 |
CN110232183A (zh) * | 2018-12-07 | 2019-09-13 | 腾讯科技(深圳)有限公司 | 关键词提取模型训练方法、关键词提取方法、装置及存储介质 |
CN110502738A (zh) * | 2018-05-18 | 2019-11-26 | 阿里巴巴集团控股有限公司 | 中文命名实体识别方法、装置、设备和查询系统 |
CN110688854A (zh) * | 2019-09-02 | 2020-01-14 | 平安科技(深圳)有限公司 | 命名实体识别方法、装置及计算机可读存储介质 |
CN111353310A (zh) * | 2020-02-28 | 2020-06-30 | 腾讯科技(深圳)有限公司 | 基于人工智能的命名实体识别方法、装置及电子设备 |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8316030B2 (en) * | 2010-11-05 | 2012-11-20 | Nextgen Datacom, Inc. | Method and system for document classification or search using discrete words |
US9087297B1 (en) * | 2010-12-17 | 2015-07-21 | Google Inc. | Accurate video concept recognition via classifier combination |
US9971763B2 (en) * | 2014-04-08 | 2018-05-15 | Microsoft Technology Licensing, Llc | Named entity recognition |
CN109388793B (zh) * | 2017-08-03 | 2023-04-07 | 阿里巴巴集团控股有限公司 | 实体标注方法、意图识别方法及对应装置、计算机存储介质 |
US10482179B2 (en) * | 2017-10-09 | 2019-11-19 | Raytheon Company | Requirement document language quality feedback and improvement |
RU2679988C1 (ru) * | 2017-12-11 | 2019-02-14 | Общество с ограниченной ответственностью "Аби Продакшн" | Извлечение информационных объектов с помощью комбинации классификаторов |
US11010561B2 (en) * | 2018-09-27 | 2021-05-18 | Apple Inc. | Sentiment prediction from textual data |
CN110209812B (zh) * | 2019-05-07 | 2022-04-22 | 北京地平线机器人技术研发有限公司 | 文本分类方法和装置 |
CN110569366B (zh) * | 2019-09-09 | 2023-05-23 | 腾讯科技(深圳)有限公司 | 文本的实体关系抽取方法、装置及存储介质 |
US11095600B2 (en) * | 2019-12-09 | 2021-08-17 | Oracle International Corporation | End-to-end email tag prediction |
-
2020
- 2020-02-28 CN CN202010127101.0A patent/CN111353310B/zh active Active
- 2020-11-10 WO PCT/CN2020/127737 patent/WO2021169400A1/zh active Application Filing
-
2022
- 2022-03-02 US US17/685,283 patent/US20220188521A1/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106547733A (zh) * | 2016-10-19 | 2017-03-29 | 中国国防科技信息中心 | 一种面向特定文本的命名实体识别方法 |
CN110502738A (zh) * | 2018-05-18 | 2019-11-26 | 阿里巴巴集团控股有限公司 | 中文命名实体识别方法、装置、设备和查询系统 |
CN109101481A (zh) * | 2018-06-25 | 2018-12-28 | 北京奇艺世纪科技有限公司 | 一种命名实体识别方法、装置及电子设备 |
CN110232183A (zh) * | 2018-12-07 | 2019-09-13 | 腾讯科技(深圳)有限公司 | 关键词提取模型训练方法、关键词提取方法、装置及存储介质 |
CN110688854A (zh) * | 2019-09-02 | 2020-01-14 | 平安科技(深圳)有限公司 | 命名实体识别方法、装置及计算机可读存储介质 |
CN111353310A (zh) * | 2020-02-28 | 2020-06-30 | 腾讯科技(深圳)有限公司 | 基于人工智能的命名实体识别方法、装置及电子设备 |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114036933A (zh) * | 2022-01-10 | 2022-02-11 | 湖南工商大学 | 基于法律文书的信息抽取方法 |
CN114048293A (zh) * | 2022-01-11 | 2022-02-15 | 广东拓思软件科学园有限公司 | 一种缺陷报告融合方法、装置、电子设备及存储介质 |
CN114841163A (zh) * | 2022-03-31 | 2022-08-02 | 阿里巴巴(中国)有限公司 | 实体识别方法、装置、设备及存储介质 |
CN116822632A (zh) * | 2023-08-28 | 2023-09-29 | 腾讯科技(深圳)有限公司 | 一种文本数据的推理方法、装置、存储介质和电子设备 |
CN116822632B (zh) * | 2023-08-28 | 2024-01-05 | 腾讯科技(深圳)有限公司 | 一种文本数据的推理方法、装置、存储介质和电子设备 |
Also Published As
Publication number | Publication date |
---|---|
CN111353310A (zh) | 2020-06-30 |
CN111353310B (zh) | 2023-08-11 |
US20220188521A1 (en) | 2022-06-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021169400A1 (zh) | 基于人工智能的命名实体识别方法、装置及电子设备 | |
US10534863B2 (en) | Systems and methods for automatic semantic token tagging | |
WO2022007823A1 (zh) | 一种文本数据处理方法及装置 | |
WO2020237856A1 (zh) | 基于知识图谱的智能问答方法、装置及计算机存储介质 | |
WO2023065211A1 (zh) | 一种信息获取方法以及装置 | |
US8577938B2 (en) | Data mapping acceleration | |
KR102179890B1 (ko) | 텍스트 데이터 수집 및 분석을 위한 시스템 | |
CN111539197A (zh) | 文本匹配方法和装置以及计算机系统和可读存储介质 | |
CN107844533A (zh) | 一种智能问答系统及分析方法 | |
WO2024098524A1 (zh) | 文本视频的互检索及模型训练方法、装置、设备及介质 | |
CN108763202A (zh) | 识别敏感文本的方法、装置、设备及可读存储介质 | |
KR20150041908A (ko) | 정답 유형 자동 분류 방법 및 장치, 이를 이용한 질의 응답 시스템 | |
CN113704420A (zh) | 文本中的角色识别方法、装置、电子设备及存储介质 | |
CN116975199A (zh) | 一种文本预测方法、装置、设备和存储介质 | |
KR20240020166A (ko) | Esg 보조 툴을 이용하여 정형화된 esg 데이터로 기계학습 모델을 학습하는 방법 및 기계학습 모델로 자동완성된 esg 문서를 생성하는 서비스 서버 | |
CN115730597A (zh) | 多级语义意图识别方法及其相关设备 | |
CN114911915A (zh) | 一种基于知识图谱的问答搜索方法、系统、设备和介质 | |
CN114841138A (zh) | 行间的机器阅读 | |
CN112487154B (zh) | 一种基于自然语言的智能搜索方法 | |
US11501071B2 (en) | Word and image relationships in combined vector space | |
CN112925983A (zh) | 一种电网资讯信息的推荐方法及系统 | |
US20240095445A1 (en) | Systems and methods for language modeling with textual clincal data | |
CN114491076B (zh) | 基于领域知识图谱的数据增强方法、装置、设备及介质 | |
CN112347289B (zh) | 一种图像管理方法及终端 | |
CN118535715B (zh) | 一种基于树状结构知识库的自动回复方法、设备及存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20921215 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20921215 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 16.02.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20921215 Country of ref document: EP Kind code of ref document: A1 |