CN116842152A - Science and technology policy question-answering method and device for fine-tuning language big model - Google Patents

Science and technology policy question-answering method and device for fine-tuning language big model Download PDF

Info

Publication number
CN116842152A
CN116842152A CN202310681071.1A CN202310681071A CN116842152A CN 116842152 A CN116842152 A CN 116842152A CN 202310681071 A CN202310681071 A CN 202310681071A CN 116842152 A CN116842152 A CN 116842152A
Authority
CN
China
Prior art keywords
policy
document
model
language
technical
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310681071.1A
Other languages
Chinese (zh)
Inventor
林卓
郑照洲
张福山
方延风
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fujian Science And Technology Information Research Institute (fujian Productivity Promotion Center)
Original Assignee
Fujian Science And Technology Information Research Institute (fujian Productivity Promotion Center)
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fujian Science And Technology Information Research Institute (fujian Productivity Promotion Center) filed Critical Fujian Science And Technology Information Research Institute (fujian Productivity Promotion Center)
Priority to CN202310681071.1A priority Critical patent/CN116842152A/en
Publication of CN116842152A publication Critical patent/CN116842152A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation or dialogue systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/14Image acquisition
    • G06V30/148Segmentation of character regions
    • G06V30/153Segmentation of character regions using recognition of characters or words

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Human Computer Interaction (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The application discloses a technological policy question-answering method and device for fine-tuning a language model, and relates to the technical field of language models. The method comprises the following steps: collecting questions and answers from the technical policy field data, and constructing a question-answer data set of the technical policy field; saving the technical policy rules as a text data file; fine tuning the pre-trained language big model by using a question-answer data set in the technical policy field; analyzing and segmenting a text data file of the technical policy regulation, and storing segmented document vectors into a database for supporting vector retrieval; when the user problem is acquired, at least one section of document vector matched with the problem is acquired through a database for supporting vector retrieval, then the acquired document vector is used as a Prompt, and a trimmed language large model is input to obtain a model output result. The application realizes a more accurate, efficient and highly-adaptive technological policy question-answering system and provides more reliable support for technological policy decision and consultation.

Description

Science and technology policy question-answering method and device for fine-tuning language big model
Technical Field
The application relates to the technical field of language models, in particular to a technological policy question-answering method and device for fine-tuning a language model.
Background
Currently, large language models have made important breakthroughs in the field of natural language processing, such as GPT-3, GLM, and the like. The models can generate high-quality texts, and have wide application prospects. However, these models have some drawbacks in dealing with problems in certain areas. In particular, when applied to the technical policy field, existing language big models often lack understanding of the expertise and complexity, resulting in poor effectiveness in solving the technical policy problem, easy misleading answers or lack of confidence.
Disclosure of Invention
The technical problem to be solved by the application is to provide a technological policy question-answering method and a technological policy question-answering device for fine-tuning a language big model, which uses a question-answering data set in the technological policy field to fine-tune a pre-trained language big model, and uses vectorization engineering to enable the language big model to have higher accuracy and reliability in question-answering tasks in the technological policy field.
In a first aspect, the present application provides a technological policy question-answering method for fine-tuning a language big model, including:
technology policy data set construction: collecting questions and answers from the technical policy field data, and constructing a question-answer data set of the technical policy field; saving the technical policy rules as a text data file;
language big model fine tuning: fine tuning a pre-trained language large model GLM by using a question-answer data set in the technical policy field, and simulating the updating quantity of parameters by adding a bypass beside an original model and low-rank decomposition;
vectorization engineering: analyzing and segmenting a text data file of the technical policy regulation, and storing segmented document vectors into a database for supporting vector retrieval;
answer output: when the user problem is acquired, at least one section of document vector matched with the problem is acquired through a database for supporting vector retrieval, the acquired document vector is used as a Prompt, a trimmed language big model is input, and a model output result is output as an answer.
Further, the language big model fine tuning specifically includes:
let the pre-trained matrix be W 0 ∈R d*k Its update is expressed as:
W 0 +ΔW=W 0 +BA,B∈R d*r ,A∈R r*k
where d, k represents the dimension of the matrix, rank r<<min (d, k), R represents a real set, W 0 Representing a pre-training matrix, the training method is as follows:
during training, the original model is fixed, and only the dimension-reducing matrix A and the dimension-increasing matrix B are trained;
during reasoning, BA is added to the original parameters, and no extra reasoning delay is introduced;
initializing, wherein A adopts Gaussian distribution initialization, B initializes to all 0, and a bypass is ensured to be a 0 matrix at the beginning of training;
rank selection: for general tasks, r=any of 1,2,4,8 values, and for some widely divergent tasks, a larger r is chosen.
Further, the vectorization process specifically includes the following steps:
loading a document: analyzing text data files of technical policy regulations, and converting all the grabbed policy files into a TXT text;
splitting a document: cutting TXT text into document vectors, wherein the length of the document vectors does not exceed a specified value, and the last character is a specified separator;
storing and retrieving document vectors: the document vectors are stored in a database supporting vector retrieval for matching the most similar document vectors from among the document vectors stored in the database after calculating the text vectors from the input text.
Further, all policy files captured are converted into one TXT text using LangChain library.
Further, the database supporting vector retrieval is a vector store.
In a second aspect, the present application provides a technological policy question-answering apparatus for fine-tuning a language big model, comprising:
the science and technology policy data set construction module: collecting questions and answers from the technical policy field data, and constructing a question-answer data set of the technical policy field; saving the technical policy rules as a text data file;
language big model fine tuning module: fine tuning a pre-trained language large model GLM by using a question-answer data set in the technical policy field, and simulating the updating quantity of parameters by adding a bypass beside an original model and low-rank decomposition;
the vectorization engineering module: analyzing and segmenting a text data file of the technical policy regulation, and storing segmented document vectors into a database for supporting vector retrieval;
answer output module: when the user problem is acquired, at least one section of document vector matched with the problem is acquired through a database for supporting vector retrieval, the acquired document vector is used as a Prompt, a trimmed language big model is input, and a model output result is output as an answer.
Further, the language big model fine tuning module is specifically configured to perform:
let the pre-trained matrix be W 0 ∈R d*k Its update is expressed as:
W 0 +ΔW=W 0 +BA,B∈R d*r ,A∈R r*k
where d, k represents the dimension of the matrix, rank r<<min (d, k), R represents a real set, W 0 Representing a pre-training matrix, the training method is as follows:
during training, the original model is fixed, and only the dimension-reducing matrix A and the dimension-increasing matrix B are trained;
during reasoning, BA is added to the original parameters, and no extra reasoning delay is introduced;
initializing, wherein A adopts Gaussian distribution initialization, B initializes to all 0, and a bypass is ensured to be a 0 matrix at the beginning of training;
rank selection: for general tasks, r=any of 1,2,4,8 values, and for some widely divergent tasks, a larger r is chosen.
Further, the vectorization module is specifically configured to perform the following steps:
loading a document: analyzing text data files of technical policy regulations, and converting all the grabbed policy files into a TXT text;
splitting a document: cutting TXT text into document vectors, wherein the length of the document vectors does not exceed a specified value, and the last character is a specified separator;
storing and retrieving document vectors: the document vectors are stored in a database supporting vector retrieval for matching the most similar document vectors from among the document vectors stored in the database after calculating the text vectors from the input text.
Further, all policy files captured are converted into one TXT text using LangChain library.
Further, the database supporting vector retrieval is a vector store.
One or more technical solutions provided in the embodiments of the present application at least have the following technical effects or advantages:
the application effect of the language big model in the technical policy field can be improved through improvement points such as the construction of the technical policy data set, the fine adjustment method of the model, the injection of field knowledge and the like, the input limitation of the model is solved through vectorization and segmentation, and finally the answer accuracy is greatly improved through the Prompt engineering. By the method, a more accurate, efficient and highly-adaptive technological policy question-answering system can be realized, and more reliable support is provided for technological policy decision and consultation.
The foregoing description is only an overview of the present application, and is intended to be implemented in accordance with the teachings of the present application in order that the same may be more clearly understood and to make the same and other objects, features and advantages of the present application more readily apparent.
Drawings
The application will be further described with reference to examples of embodiments with reference to the accompanying drawings.
FIG. 1 is a flow chart of a method according to a first embodiment of the application;
fig. 2 is a schematic structural diagram of a device in a second embodiment of the present application.
Detailed Description
The embodiment of the application provides a technological policy question-answering method and device for fine-tuning a language big model, which are based on the trained language big model, and combines the structure and content characteristics of documents to realize automatic interpretation and knowledge extraction of scientific research documents, thereby improving the accuracy of interpretation and understanding.
The technical scheme in the embodiment of the application has the following overall thought:
data set construction: questions and answers are collected from relevant documents, policy documents and expert knowledge in the technical policy field, a question-answer data set specific to the technical policy field is constructed, and technical policy regulations are stored as text documents.
Fine tuning of the pre-training model: by introducing domain knowledge and data related to the technical policy in the pre-training process and fine-tuning specific task data of the domain, the language model has better technical policy understanding capability.
Vectorization engineering: aiming at scientific and technological data in a certain field, the input limit of a model is solved through vectorization and segmentation by collecting files, and finally, the answer accuracy is greatly improved through a Prompt project. .
Example 1
The embodiment provides a technological policy question-answering method for fine-tuning a language big model, as shown in fig. 1, which may include the following procedures:
technology policy data set construction: questions and answers are collected from technical policy domain data (e.g., domain knowledge, policy rules, related terms, and practice experience related to a technical policy, etc.), a question-answer data set of the technical policy domain is constructed, the data format may be { q, a }, q represents the collected technical policy questions, and a represents answers to the technical policy questions. The technological policy rules are stored as text data files, all policy rules can be collected into one folder, and the storage formats are pdf and txt.
Language big model fine tuning: the pre-trained language big model GLM is fine-tuned by using a question-answer data set in the technical policy field, and the update amount of parameters is simulated by adding a bypass beside the original model and low-rank decomposition. The process is as follows:
let the pre-trained matrix be W 0 ∈R d*k Its update is expressed as:
W 0 +ΔW=W 0 +BA,B∈R d*r ,A∈R r*k
where d, k represents the dimension of the matrix, rank r<<min (d, k), R represents a real set, W 0 Representing a pre-training matrix, the training method is as follows:
during training, the original model is fixed, and only the dimension-reducing matrix A and the dimension-increasing matrix B are trained;
during reasoning, BA is added to the original parameters, and no extra reasoning delay is introduced;
initializing, wherein A adopts Gaussian distribution initialization, B initializes to all 0, and a bypass is ensured to be a 0 matrix at the beginning of training;
rank selection: for general tasks, r=any of 1,2,4,8 values, and for some widely divergent tasks, a larger r is chosen. R=8 in this embodiment.
Vectorization engineering: analyzing and segmenting a text data file of the technical policy regulation, and storing segmented document vectors into a database for supporting vector retrieval; the process is as follows:
loading a document: analyzing text data files of technical policy regulations, and converting all the grabbed policy files into a TXT text by using a Langchain library;
splitting a document: cutting TXT text into document vectors, wherein the length of the document vectors does not exceed a specified value (in order to meet the limitation of the input length of a model, such as 2048token is not exceeded), and the last characters are specified separators (such as line feed, period, and the like), so that the integrity of semantics is ensured;
storing and retrieving document vectors: the document vectors are stored in a database (e.g., vector store) supporting vector retrieval for matching the most similar document vectors from among the document vectors stored in the database after calculating the text vectors from the input text, i.e., for finding several most relevant documents based on user questions.
Answer output: when the user problem is acquired, at least one section of document vector matched with the problem is acquired through a database for supporting vector retrieval, the acquired document vector is used as a Prompt, a language big model after fine adjustment is input, and a model output result is used as an answer to be output, so that a more accurate result is obtained.
Based on the same inventive concept, the application also provides a device corresponding to the method in the first embodiment, and the details of the second embodiment are shown.
Example two
In this embodiment, a technological policy question-answering device for fine-tuning a language big model is provided, as shown in fig. 2, including:
the science and technology policy data set construction module: collecting questions and answers from the technical policy field data, and constructing a question-answer data set of the technical policy field; saving the technical policy rules as a text data file;
language big model fine tuning module: fine tuning a pre-trained language large model GLM by using a question-answer data set in the technical policy field, and simulating the updating quantity of parameters by adding a bypass beside an original model and low-rank decomposition;
the language big model fine tuning module is specifically configured to execute:
let the pre-trained matrix be W 0 ∈R d*k Its update is expressed as:
W 0 +ΔW=W 0 +BA,B∈R d*r ,A∈R r*k
where d, k represents the dimension of the matrix, rank r<<min (d, k), R represents a real set, W 0 Representing a pre-training matrix, the training method is as follows:
during training, the original model is fixed, and only the dimension-reducing matrix A and the dimension-increasing matrix B are trained;
during reasoning, BA is added to the original parameters, and no extra reasoning delay is introduced;
initializing, wherein A adopts Gaussian distribution initialization, B initializes to all 0, and a bypass is ensured to be a 0 matrix at the beginning of training;
rank selection: for general tasks, r=any of 1,2,4,8 values, and for some widely divergent tasks, a larger r is chosen.
The vectorization engineering module: analyzing and segmenting a text data file of the technical policy regulation, and storing segmented document vectors into a database for supporting vector retrieval;
the vectorization module is specifically configured to perform the following steps:
loading a document: analyzing text data files of technical policy regulations, and converting all the grabbed policy files into a TXT text by using a Langchain library;
splitting a document: cutting TXT text into document vectors, wherein the length of the document vectors does not exceed a specified value, and the last character is a specified separator;
storing and retrieving document vectors: the document vectors are stored in a database (e.g., vector store) supporting vector retrieval for matching the most similar document vector from among the document vectors stored in the database after computing the text vectors from the input text.
Answer output module: when the user problem is acquired, at least one section of document vector matched with the problem is acquired through a database for supporting vector retrieval, the acquired document vector is used as a Prompt, a trimmed language big model is input, and a model output result is output as an answer.
The application can improve the application effect of the language large model in the technical policy field through data set construction, fine tuning method, field knowledge injection and other improvement points, solves the model input limitation through vectorization and segmentation modes, and finally greatly improves the answer accuracy through the Prompt engineering. By the method, a more accurate, efficient and highly-adaptive technological policy question-answering system can be realized, and more reliable support is provided for technological policy decision and consultation. The method has the following advantages of 1 and high accuracy: by fine tuning the data set in the technical policy field, the language big model has higher accuracy and reliability in question-answering tasks in the field. 2. High efficiency: through injecting domain-specific knowledge, the language big model can better understand and answer questions related to technical policies, and the question and answer efficiency is improved. 3. High adaptability: the method can be suitable for different types of technological policy question-answering tasks, and has certain universality and flexibility. 4. Providing domain-specific support: practitioners in the technical policy field can rapidly acquire accurate technical policy related information by using the method of the application, thereby improving the working efficiency and the decision quality.
It will be apparent to those skilled in the art that embodiments of the present application may be provided as a method, apparatus, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
While specific embodiments of the application have been described above, it will be appreciated by those skilled in the art that the specific embodiments described are illustrative only and not intended to limit the scope of the application, and that equivalent modifications and variations of the application in light of the spirit of the application will be covered by the claims of the present application.

Claims (10)

1. A technological policy question-answering method for fine-tuning a language big model, comprising:
technology policy data set construction: collecting questions and answers from the technical policy field data, and constructing a question-answer data set of the technical policy field; saving the technical policy rules as a text data file;
language big model fine tuning: fine tuning a pre-trained language large model GLM by using a question-answer data set in the technical policy field, and simulating the updating quantity of parameters by adding a bypass beside an original model and low-rank decomposition;
vectorization engineering: analyzing and segmenting a text data file of the technical policy regulation, and storing segmented document vectors into a database for supporting vector retrieval;
answer output: when the user problem is acquired, at least one section of document vector matched with the problem is acquired through a database for supporting vector retrieval, the acquired document vector is used as a Prompt, a trimmed language big model is input, and a model output result is output as an answer.
2. The method according to claim 1, wherein the language big model fine-tuning specifically comprises:
let the pre-trained matrix be W 0 ∈R d*k Its update is expressed as:
W 0 +ΔW=W 0 +BA,B∈R d*r ,A∈R r*k
where d, k represents the dimension of the matrix, rank r<<min (d, k), R represents a real set, W 0 Representing a pre-training matrix, the training method is as follows:
during training, the original model is fixed, and only the dimension-reducing matrix A and the dimension-increasing matrix B are trained;
during reasoning, BA is added to the original parameters, and no extra reasoning delay is introduced;
initializing, wherein A adopts Gaussian distribution initialization, B initializes to all 0, and a bypass is ensured to be a 0 matrix at the beginning of training;
rank selection: for general tasks, r=any of 1,2,4,8 values, and for some widely divergent tasks, a larger r is chosen.
3. The method according to claim 1 or 2, characterized in that the vectorization step comprises in particular the following steps:
loading a document: analyzing text data files of technical policy regulations, and converting all the grabbed policy files into a TXT text;
splitting a document: cutting TXT text into document vectors, wherein the length of the document vectors does not exceed a specified value, and the last character is a specified separator;
storing and retrieving document vectors: the document vectors are stored in a database supporting vector retrieval for matching the most similar document vectors from among the document vectors stored in the database after calculating the text vectors from the input text.
4. A method according to claim 3, characterized in that: all policy files fetched are converted into one TXT text using LangChain library.
5. The method according to claim 1, characterized in that: the database supporting vector retrieval is a vector store.
6. A technological policy question-answering apparatus for fine-tuning a language model, comprising:
the science and technology policy data set construction module: collecting questions and answers from the technical policy field data, and constructing a question-answer data set of the technical policy field; saving the technical policy rules as a text data file;
language big model fine tuning module: fine tuning a pre-trained language large model GLM by using a question-answer data set in the technical policy field, and simulating the updating quantity of parameters by adding a bypass beside an original model and low-rank decomposition;
the vectorization engineering module: analyzing and segmenting a text data file of the technical policy regulation, and storing segmented document vectors into a database for supporting vector retrieval;
answer output module: when the user problem is acquired, at least one section of document vector matched with the problem is acquired through a database for supporting vector retrieval, the acquired document vector is used as a Prompt, a trimmed language big model is input, and a model output result is output as an answer.
7. The apparatus according to claim 6, wherein the language big model fine tuning module is specifically configured to perform:
let the pre-trained matrix be W 0 ∈R d*k Its update is expressed as:
W 0 +ΔW=W 0 +BA,B∈R d*r ,A∈R r*k
where d, k represents the dimension of the matrix, rank r<<min (d, k), R represents a real set, W 0 Representing a pre-training matrix, the training method is as follows:
during training, the original model is fixed, and only the dimension-reducing matrix A and the dimension-increasing matrix B are trained;
during reasoning, BA is added to the original parameters, and no extra reasoning delay is introduced;
initializing, wherein A adopts Gaussian distribution initialization, B initializes to all 0, and a bypass is ensured to be a 0 matrix at the beginning of training;
rank selection: for general tasks, r=any of 1,2,4,8 values, and for some widely divergent tasks, a larger r is chosen.
8. The apparatus according to claim 6 or 7, wherein the vectorization module is specifically configured to perform the following steps:
loading a document: analyzing text data files of technical policy regulations, and converting all the grabbed policy files into a TXT text;
splitting a document: cutting TXT text into document vectors, wherein the length of the document vectors does not exceed a specified value, and the last character is a specified separator;
storing and retrieving document vectors: the document vectors are stored in a database supporting vector retrieval for matching the most similar document vectors from among the document vectors stored in the database after calculating the text vectors from the input text.
9. The apparatus according to claim 8, wherein: all policy files fetched are converted into one TXT text using LangChain library.
10. The apparatus according to claim 6, wherein: the database supporting vector retrieval is a vector store.
CN202310681071.1A 2023-06-09 2023-06-09 Science and technology policy question-answering method and device for fine-tuning language big model Pending CN116842152A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310681071.1A CN116842152A (en) 2023-06-09 2023-06-09 Science and technology policy question-answering method and device for fine-tuning language big model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310681071.1A CN116842152A (en) 2023-06-09 2023-06-09 Science and technology policy question-answering method and device for fine-tuning language big model

Publications (1)

Publication Number Publication Date
CN116842152A true CN116842152A (en) 2023-10-03

Family

ID=88171607

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310681071.1A Pending CN116842152A (en) 2023-06-09 2023-06-09 Science and technology policy question-answering method and device for fine-tuning language big model

Country Status (1)

Country Link
CN (1) CN116842152A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117235243A (en) * 2023-11-16 2023-12-15 青岛民航凯亚系统集成有限公司 Training optimization method for large language model of civil airport and comprehensive service platform
CN117633179A (en) * 2023-11-30 2024-03-01 武汉铁路职业技术学院 Intelligent question-answering assistant system for railway industry
CN117744633A (en) * 2024-02-21 2024-03-22 中国铁道科学研究院集团有限公司电子计算技术研究所 Text error correction method and device and electronic equipment
CN117807962A (en) * 2024-03-01 2024-04-02 之江实验室 Method and device for writing consultation administrative texts, storage medium and electronic equipment

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117235243A (en) * 2023-11-16 2023-12-15 青岛民航凯亚系统集成有限公司 Training optimization method for large language model of civil airport and comprehensive service platform
CN117633179A (en) * 2023-11-30 2024-03-01 武汉铁路职业技术学院 Intelligent question-answering assistant system for railway industry
CN117633179B (en) * 2023-11-30 2024-06-21 武汉铁路职业技术学院 Intelligent question-answering assistant system for railway industry
CN117744633A (en) * 2024-02-21 2024-03-22 中国铁道科学研究院集团有限公司电子计算技术研究所 Text error correction method and device and electronic equipment
CN117807962A (en) * 2024-03-01 2024-04-02 之江实验室 Method and device for writing consultation administrative texts, storage medium and electronic equipment

Similar Documents

Publication Publication Date Title
CN116842152A (en) Science and technology policy question-answering method and device for fine-tuning language big model
CN116701431A (en) Data retrieval method and system based on large language model
CN109800879B (en) Knowledge base construction method and device
CN117076693A (en) Method for constructing digital human teacher multi-mode large language model pre-training discipline corpus
US20240281659A1 (en) Augmenting machine learning language models using search engine results
CN117609444B (en) Searching question-answering method based on large model
CN111881264B (en) Method and electronic equipment for searching long text in question-answering task in open field
CN111552773A (en) Method and system for searching key sentence of question or not in reading and understanding task
CN117668181A (en) Information processing method, device, terminal equipment and storage medium
CN113590811A (en) Text abstract generation method and device, electronic equipment and storage medium
CN106776782B (en) Semantic similarity obtaining method and device based on artificial intelligence
CN117708282A (en) Knowledge question-answering method and system based on large language model
CN110674283B (en) Intelligent extraction method and device for text abstracts, computer equipment and storage medium
CN116680387A (en) Dialogue reply method, device, equipment and storage medium based on retrieval enhancement
CN111126066B (en) Method and device for determining Chinese congratulation technique based on neural network
CN114756617A (en) Method, system, equipment and storage medium for extracting structured data of engineering archives
CN114595684A (en) Abstract generation method and device, electronic equipment and storage medium
CN109388786B (en) Document similarity calculation method, device, equipment and medium
CN112036183A (en) Word segmentation method and device based on BilSTM network model and CRF model, computer device and computer storage medium
CN118396123B (en) Text generation method, apparatus, computer program product, electronic device, and medium
CN117033733B (en) Intelligent automatic classification and label generation system and method for library resources
CN114896362A (en) Method and device for analyzing similarity of patent texts, storage medium and equipment
CN118427329A (en) Method and system for processing public accumulation policy question-answer data
CN118503376A (en) Data processing method and device for multi-hop problem
CN117851574A (en) Text dialogue method and device and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination