WO2021135091A1 - 一种基于深度学习的目标软文的生成方法及装置 - Google Patents
一种基于深度学习的目标软文的生成方法及装置 Download PDFInfo
- Publication number
- WO2021135091A1 WO2021135091A1 PCT/CN2020/097007 CN2020097007W WO2021135091A1 WO 2021135091 A1 WO2021135091 A1 WO 2021135091A1 CN 2020097007 W CN2020097007 W CN 2020097007W WO 2021135091 A1 WO2021135091 A1 WO 2021135091A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- target
- title
- word segmentation
- titles
- input
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/166—Editing, e.g. inserting or deleting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
Definitions
- the invention relates to the technical field of natural language processing, in particular to a method and device for generating target soft text based on deep learning.
- Marketing essays are often used when new products are promoted in the market.
- Marketing essays usually consist of three parts: title, introduction, and marketing text.
- title uses vivid and concise language to indicate the product to be marketed, which is spectacular.
- the introduction plays a guiding role and guides the direction of consumption. It leads to the following marketing text, and the marketing text introduces the product and recommends marketing.
- the embodiments of the present invention provide a method and device for generating target soft texts based on deep learning to overcome the low production efficiency of manually writing target soft texts in the prior art, and the template-generated target soft text sentence pattern is fixed, Problems such as dullness and insufficient diversification.
- the technical solution adopted by the present invention is:
- a method for generating target soft articles based on deep learning includes the following steps:
- the target title, the target lead, and the target text are assembled to obtain multiple target soft articles.
- the input information is input into the second generation model to generate at least one target text.
- the method also includes a process of building a title library, including:
- the second word segmentation result and the first keyword are input into a third generation model to obtain a plurality of new titles, and the title library is composed of the new titles.
- construction process of the title library further includes:
- a third generation model is trained based on a preset algorithm.
- the method further includes a construction process of the first generative model, including:
- the title matching module is used to receive related information of the target object, and match several adapted target titles from the title library according to the related information.
- the titles in the title library are expanded by the collected titles through the third generation model Come
- the lead generation module is used to input the target title into the first generation model to generate at least one target lead;
- a text generation module configured to generate at least one input information conforming to a preset structure according to the related information and preset rules, input the input information into the second generation model, and generate at least one target text;
- the information assembly module assembles the target title, the target lead, and the target text to obtain multiple target soft texts.
- the text generation module includes:
- the first word segmentation unit is configured to perform word segmentation processing on the related information, and extract a target word segmentation that meets a preset condition from the obtained first word segmentation result;
- the word segmentation reorganization unit is used to reorganize the target word segmentation to obtain at least one piece of input information that conforms to a preset structure
- the text generation unit is used to input the input information into the second generation model to generate at least one target text.
- the device further includes a first building module, including:
- the second word segmentation unit is used to perform word segmentation processing on several collected first sample titles to obtain the second word segmentation result;
- the first extraction unit is configured to extract the first keyword from the first sample title by using a preset first keyword extraction method
- the title generation unit is configured to input the second word segmentation result and the first keyword into a third generation model to obtain a plurality of new titles, and the title library is composed of the new titles.
- the first building module further includes:
- a first intersection unit configured to intersect the first keyword set and the second word segmentation result to obtain an input data set
- the first training unit is configured to take the data of the input data set as input and the target title as output, and train a third generation model based on a preset algorithm.
- the device further includes a second building module, including:
- the third word segmentation unit is used to perform word segmentation processing on several collected second sample titles and introductory pairs corresponding to the second sample titles;
- the second extraction unit is configured to extract the second keyword from the second sample title by using a preset second keyword extraction method
- the second intersection unit is used to take the intersection of the second keyword set and the second sample title after each word segmentation to obtain the target keyword;
- the lead expansion unit is used to traverse each of the second sample titles, match the target keywords with the lead that corresponds to the second sample title after full word segmentation, and obtain the successfully matched lead as the current second sample The new lead of the title;
- the second training unit is configured to take the second sample title as input, the lead corresponding to the second sample title and the new lead as output, and train a first generative model based on a preset algorithm.
- the method and device for generating target soft text based on deep learning receive relevant information of the target object, and according to the relevant information, match several suitable target titles from the title library, and the titles in the title library
- the collected headline is extended by the third generation model, the target headline is input into the first generation model, at least one target lead is generated, and at least one input information conforming to the preset structure is generated according to related information and preset rules.
- Input information into the second generative model generate at least one target text, assemble the target title, target introduction, and target text to obtain multiple target texts.
- deep learning and natural language processing technology it can realize the automatic intelligence of marketing texts. Diversified generation, saving the investment of operating personnel, improving the production efficiency of marketing soft text, effectively avoiding the problem of low handwriting efficiency, and avoiding the dull problem of template generation;
- the method and device for generating target soft text based on deep learning obtain the second word segmentation result by performing word segmentation processing on a number of collected first sample titles, and use the preset first keyword extraction
- the method extracts the first keyword from the first sample title, inputs the second word segmentation result and the first keyword into the third generation model, obtains multiple new titles, and utilizes the existing limited titles Expand the number of titles in the title library;
- the method and device for generating target soft texts based on deep learning perform word segmentation processing on a number of collected second sample titles and introductory pairs corresponding to the second sample titles, using a preset
- the second keyword extraction method extracts the second keyword from the second sample title, takes the intersection of the second keyword set and the second sample title after each word segmentation, obtains the target keyword, and traverses For each of the second sample titles, match the target keyword with the lead corresponding to the second sample title after full word segmentation, and obtain the successfully matched lead as the new lead of the current second sample heading, and
- the second sample title is used as input, and the introduction corresponding to the second sample title and the new introduction are used as output.
- the first generation model is trained based on a preset algorithm, which expands the training data of the introduction generation model and avoids Due to insufficient training data, it is easy to cause problems such as over-fitting and poor generation effect.
- Fig. 1 is a flowchart showing a method for generating a target soft article based on deep learning according to an exemplary embodiment
- Fig. 2 is a flowchart of generating at least one input information conforming to a preset structure according to related information and preset rules, inputting the input information into a second generation model, and generating at least one target text according to an exemplary embodiment;
- Fig. 3 is a flowchart showing a construction process of a title library according to an exemplary embodiment
- Fig. 4 is a flowchart showing a construction process of a title library according to another exemplary embodiment
- Fig. 5 is a flowchart showing a construction process of a first generation model according to an exemplary embodiment
- Fig. 6 is a schematic structural diagram of a device for generating target soft text based on deep learning according to an exemplary embodiment.
- the method for generating target soft text based on deep learning provided by the present invention firstly retrieves the adapted title from the title database according to the relevant information of the target object, and then generates the introductory and marketing language in turn according to the matched title and related information (ie marketing Body), finally assemble the target title, introduction and marketing body, and output multiple marketing soft articles.
- the Seq2Seq algorithm is used to generate the introduction and the marketing text, which can effectively avoid the problem of low handwriting efficiency and at the same time avoid the dullness of template generation.
- Seq2Seq is a generative architecture composed of an encoder and a decoder. It generates an output sequence Y according to the input sequence X. It is widely used in tasks such as translation, automatic text summarization, and automated robot question answering.
- Fig. 1 is a flowchart showing a method for generating a target soft text based on deep learning according to an exemplary embodiment. Referring to Fig. 1, the method includes the following steps:
- S1 Receive related information of the target object, and match several adapted target titles from the title library according to the related information, and the titles in the title library are expanded from the collected titles through the third generation model.
- the target soft article generally contains three parts: title, introduction and body.
- the target soft essay in the embodiment of the present invention includes marketing soft essays. Taking marketing soft essays as an example, the marketing soft essays include three parts: title, introduction, and body.
- the relevant information of the target object in the embodiment of the present invention includes the title of the product for which the target soft text is to be generated, or the description information of the target object for which the target soft text is to be generated, and in the embodiment of the present invention, the received relevant information may be user input , And the relevant information entered by the user can be one or more titles of a certain category of products.
- the title matching method After receiving the relevant information of the target object input by the user, according to the preset title matching method (for example, character string matching after word segmentation, similarity matching, etc.), from the title library, a number of matching information is matched.
- the target title wherein the title in the title library is expanded from the collected title through the third generation model.
- the title matching method is not specifically limited, and the user can set it according to specific needs.
- S2 Input the target title into the first generation model to generate at least one target lead.
- the first generation model is a natural language processing model pre-trained by using a preset algorithm (for example, the Seq2Seq algorithm).
- the input of the model is the above-mentioned target title, and the output is the target introduction corresponding to the target title.
- the number of target slogans output by the first generation model can be one or more, and there is no limitation here.
- S3 Generate at least one input information conforming to the preset structure according to the related information and preset rules, and input the input information into the second generation model to generate at least one target text.
- the second generation model is also a natural language processing model pre-trained by using a preset algorithm (for example, the Seq2Seq algorithm).
- a preset algorithm for example, the Seq2Seq algorithm.
- the input of the second generation model is expanded. Therefore, before generating the target text, first generate at least one input information conforming to the preset structure according to related information and preset rules, and then input the acquired input information into the second generation model to generate at least one target text.
- at least one means that there can be one or more.
- At least one piece of input information conforming to a preset structure is generated according to the related information and a preset rule, and the input information is input to
- generating at least one target text includes:
- S101 Perform word segmentation processing on the related information, and extract a target word segmentation meeting a preset condition from the obtained first word segmentation result.
- the relevant information is mostly a structure of "modifier + category word", where the modifiers are words such as brand, function, characteristic, and material.
- the input of the second generative model is expanded by reorganizing the sequence of modifiers, so that the target text output by the second generative model can be diversified. Therefore, before generating the target text, it is necessary to perform word segmentation processing on related information to obtain the first word segmentation result, and then extract the target word segmentation that meets the preset conditions from the first word segmentation result. Since the order of modifiers is reorganized to expand the input of the second generative model, the target participle that meets the preset conditions here is the participle belonging to the modifier in the result of the first participle.
- S102 Reorganize the target word segmentation to obtain at least one piece of input information that conforms to a preset structure.
- a reorganization mechanism can be preset according to actual needs, for example, the order of modifiers after regrouping words. Then, according to the reorganization mechanism, the target word segmentation obtained in the above steps is reorganized, and a plurality of input information conforming to the preset structure is output.
- the preset structure can be a "modifier + category word" structure, and the user can set and adjust it according to actual needs, and there is no specific restriction here.
- S103 Input the input information into the second generation model to generate at least one target text.
- the input information obtained through the above steps is input into the second generation model to generate at least one target text.
- Fig. 3 is a flowchart showing the construction process of the title library according to an exemplary embodiment.
- the construction process of the title library includes:
- S201 Perform word segmentation processing on several collected first sample titles, and obtain a second word segmentation result.
- the adapted target title is obtained by matching from the title library according to the related information.
- the method of expanding the collected limited titles is adopted to increase the number of titles in the title library.
- S202 Use a preset first keyword extraction method to extract a first keyword from the first sample title.
- the preset first keyword extraction method is then used to extract the first keyword from the sample title, where the user can set the extraction ratio of the first keyword according to actual needs (that is, the first keyword accounts for the sample title). Proportion of the title).
- the first keyword extraction method is not specifically limited, and the user can set it according to actual needs, for example, using the TS-IDF algorithm.
- S203 Input the second word segmentation result and the first keyword into a third generation model to obtain multiple new titles, and the title library is composed of the new titles.
- the second word segmentation result and the first keyword obtained in the above steps are used as the input of the third generative model, and the output obtained (the output is the new title) is the expanded title obtained according to the target title.
- These new titles It constitutes the title library provided by the embodiment of the present invention.
- a beam search (BeamSearch) decoder can be used, so that a large number of titles can be generated.
- FIG. 4 is a flowchart of a construction process of a title library according to another exemplary embodiment.
- the construction process of the title library includes :
- S301 Perform word segmentation processing on a number of collected first sample titles to obtain a second word segmentation result
- S302 Use a preset first keyword extraction method to extract the first keyword from the sample title
- S304 Take the data of the input data set as input and the target title as output, and train a third generation model based on a preset algorithm;
- S305 Input the second word segmentation result and the first keyword into a third generation model to obtain multiple new titles, and the title library is composed of the new titles.
- the third generation model here is also a natural language processing model pre-trained by using a preset algorithm (for example, the Seq2Seq algorithm).
- a preset algorithm for example, the Seq2Seq algorithm.
- the intersection of the first keyword set and the second word segmentation result obtained in the above steps can be performed to obtain the input data set, and then the data of the input data set can be used as input.
- the title is the output, and a third generation model is trained based on a preset algorithm (for example, the Seq2Seq algorithm).
- the specific implementation process of steps S301, S302 and step S305 can refer to the specific implementation process of steps S201 to S203 described above, which will not be repeated here.
- models in different training states can be used to repeat the above steps to further expand the title.
- This method only uses existing headings (referring to the first sample headings), and uses a specific extraction method to construct input and output to train a third generative model, so that a large number of flexible headings can be obtained in a short time, saving labor costs ,Increase productivity.
- Fig. 5 is a flow chart showing the construction process of the first generative model according to an exemplary embodiment.
- the method further includes a first
- the process of building a generative model includes:
- S401 Perform word segmentation processing on several collected second sample titles and lead pairs corresponding to the second sample titles.
- the method of keyword matching is used to mine the internal relationship between the title and the lead, and one title is matched with multiple leads.
- This can greatly expand the training data of the first generation model and avoid training data. Insufficient data leads to problems such as over-fitting and poor generation effect, which effectively improves the generation effect of the first generative model.
- first collect a certain amount of title-introduction pairs in advance that is, collect a number of second sample titles and introductions corresponding to the second sample titles, and then perform the second sample title and the introduction pairs corresponding to the second sample titles Word segmentation processing, obtain its segmentation results respectively.
- S402 Use a preset second keyword extraction method to extract a second keyword from the second sample title.
- the preset second keyword extraction method is then used to extract the second keyword from the second sample title, where the user can set the extraction ratio of the second keyword according to actual needs (that is, the second keyword accounts for the sample Proportion of the title).
- the second keyword extraction method is also not specifically limited, and the user can set it according to actual needs, for example, using the TF-IDF algorithm.
- S403 Take an intersection of the second keyword set and the second sample title after each word segmentation to obtain a target keyword.
- the target keywords extracted from each second sample title in specific implementation, the second keyword set and the second sample title after each word segmentation can be intersected, and the result obtained by the intersection will be taken As the target keyword.
- an optimal matching criterion is set in advance according to actual needs, such as sorting according to the number of matched keywords, and the top 10 leading words with the largest number of matched keywords are selected as the leading words corresponding to the title.
- Traverse each second sample title use the target keywords of each second sample title to match with the lead after the full amount of word segmentation, and obtain several successfully matched leads according to the preset optimal matching criterion as the current second sample The new lead of the title, which can greatly expand the amount of data.
- S405 Take the second sample title as input, and the lead corresponding to the second sample title and the new lead as output, and train a first generation model based on a preset algorithm.
- the first generation model is also a natural language processing model pre-trained by using a preset algorithm (for example, the Seq2Seq algorithm).
- a preset algorithm for example, the Seq2Seq algorithm.
- Fig. 6 is a schematic structural diagram of an apparatus for generating target soft text based on deep learning according to an exemplary embodiment. As shown in Fig. 6, the apparatus includes:
- the title matching module is configured to receive relevant information of the target object, and match several adapted target titles from the title library according to the relevant information, and the titles in the title library are expanded from the collected titles;
- the lead generation module is used to input the target title into the first generation model to generate at least one target lead;
- a text generation module configured to generate at least one input information conforming to a preset structure according to the related information and preset rules, and input the input information into the second generation model to generate at least one target text;
- the information assembly module assembles the target title, the target introduction, and the target text to obtain multiple target soft texts.
- the text generation module includes:
- the first word segmentation unit is configured to perform word segmentation processing on the related information, and extract a target word segmentation that meets a preset condition from the obtained first word segmentation result;
- the word segmentation reorganization unit is used to reorganize the target word segmentation to obtain at least one piece of input information that conforms to a preset structure
- the text generation unit is used to input the input information into the second generation model to generate at least one target text.
- the device further includes a first building module, including:
- the second word segmentation unit is used to perform word segmentation processing on several collected first sample titles to obtain the second word segmentation result;
- the first extraction unit is configured to extract the first keyword from the first sample title by using a preset first keyword extraction method
- the title generation unit is configured to input the second word segmentation result and the first keyword into a third generation model to obtain a plurality of new titles, and the title library is composed of the new titles.
- the first building module further includes:
- a first intersection unit configured to intersect the first keyword set and the second word segmentation result to obtain an input data set
- the first training unit is configured to take the data of the input data set as input and the target title as output, and train a third generation model based on a preset algorithm.
- the device further includes a second building module, including:
- the third word segmentation unit is used to perform word segmentation processing on several collected second sample titles and introductory pairs corresponding to the second sample titles;
- the second extraction unit is configured to extract the second keyword from the second sample title by using a preset second keyword extraction method
- the second intersection unit is used to take the intersection of the second keyword set and the second sample title after each word segmentation to obtain the target keyword;
- the lead expansion unit is used to traverse each of the second sample titles, match the target keywords with the lead that corresponds to the second sample title after full word segmentation, and obtain the successfully matched lead as the current second sample The new lead of the title;
- the second training unit is configured to take the second sample title as input, the lead corresponding to the second sample title and the new lead as output, and train a first generative model based on a preset algorithm.
- the method and device for generating target soft text based on deep learning receive relevant information of the target object, and according to the relevant information, match several suitable target titles from the title library, and the titles in the title library
- the collected headline is extended by the third generation model, the target headline is input into the first generation model, at least one target lead is generated, and at least one input information conforming to the preset structure is generated according to related information and preset rules.
- Input information into the second generative model generate at least one target text, assemble the target title, target introduction, and target text to obtain multiple target texts.
- deep learning and natural language processing technology it can realize the automatic intelligence of marketing texts. Diversified generation, saving the investment of operating personnel, improving the production efficiency of marketing soft text, effectively avoiding the problem of low handwriting efficiency, and avoiding the dull problem of template generation;
- the method and device for generating target soft text based on deep learning obtain the second word segmentation result by performing word segmentation processing on a number of collected first sample titles, and use the preset first keyword extraction
- the method extracts the first keyword from the first sample title, inputs the second word segmentation result and the first keyword into the third generation model, obtains multiple new titles, and utilizes the existing limited titles Expand the number of titles in the title library;
- the method and device for generating target soft text based on deep learning are processed by segmenting a number of collected second sample titles and introductory pairs corresponding to the second sample titles, using a preset
- the second keyword extraction method extracts the second keyword from the second sample title, takes the intersection of the second keyword set and the second sample title after each word segmentation, obtains the target keyword, and traverses For each of the second sample titles, the target keywords are matched with the lead words corresponding to the second sample headings after the full word segmentation, and the successfully matched lead words are obtained as the new lead words of the current second sample headings.
- the second sample title is used as input, and the introduction corresponding to the second sample title and the new introduction are used as output.
- the first generation model is trained based on a preset algorithm, which expands the training data of the introduction generation model and avoids Due to insufficient training data, it is easy to cause problems such as over-fitting and poor generation effect.
- the device for generating a target soft text based on deep learning triggers the target soft text generation service
- only the division of the above functional modules is used as an example for illustration. In actual applications, the above functions can be changed according to needs.
- the allocation is completed by different functional modules, that is, the internal structure of the device is divided into different functional modules to complete all or part of the functions described above.
- the device for generating a target soft article based on deep learning provided by the above embodiment belongs to the same concept as the embodiment of the method for generating a target soft article based on deep learning, that is, the device is based on the method for generating a target soft article based on deep learning.
- the specific implementation process refer to the method embodiment, which will not be repeated here.
- the program can be stored in a computer-readable storage medium.
- the storage medium mentioned can be a read-only memory, a magnetic disk or an optical disk, etc.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Machine Translation (AREA)
- Character Discrimination (AREA)
Abstract
Description
Claims (10)
- 一种基于深度学习的目标软文的生成方法,其特征在于,所述方法包括如下步骤:接收目标对象的相关信息,根据所述相关信息从标题库中匹配出若干条适配的目标标题,所述标题库中的标题由采集到的标题通过第三生成模型扩展而来;将所述目标标题输入到第一生成模型中,生成至少一个目标导语;根据所述相关信息以及预设规则生成至少一个符合预设结构的输入信息,将所述输入信息输入到第二生成模型中,生成至少一个目标正文;对所述目标标题、所述目标导语以及所述目标正文进行组装,获取多篇目标软文。
- 根据权利要求1所述的基于深度学习的目标软文的生成方法,其特征在于,所述根据所述相关信息以及预设规则生成至少一个符合预设结构的输入信息,将所述输入信息输入到第二生成模型中,生成至少一个目标正文包括:对所述相关信息进行分词处理,从获取到的第一分词结果中提取出满足预设条件的目标分词;对所述目标分词进行重组,获取至少一个符合预设结构的输入信息;将所述输入信息输入到第二生成模型中,生成至少一个目标正文。
- 根据权利要求1或2所述的基于深度学习的目标软文的生成方法,其特征在于,所述方法还包括标题库的构建过程,包括:对采集到的若干第一样本标题进行分词处理,获取第二分词结果;采用预设的第一关键词提取方法从所述第一样本标题中提取出第一关键词;将所述第二分词结果以及所述第一关键词输入到第三生成模型,获取多个新的标题,所述标题库由所述新的标题构成。
- 根据权利要求3所述的基于深度学习的目标软文的生成方法,其特征在于,所述标题库的构建过程还包括:对所述第一关键词集合与所述第二分词结果取交集,获取输入数据集合;将所述输入数据集合的数据作为输入,所述目标标题作为输出,基于预设算法训练出第三生成模型。
- 根据权利要求1或2所述的基于深度学习的目标软文的生成方法,其特征在于,所述方法还包括第一生成模型的构建过程,包括:对采集到的若干第二样本标题以及与所述第二样本标题对应的导语对进行分词处理;采用预设的第二关键词提取方法从所述第二样本标题中提取出第二关键词;对所述第二关键词集合与每条分词后的所述第二样本标题取交集,获取目标关键词;遍历每一所述第二样本标题,将所述目标关键词与全量分词后的与所述第二样本标题对应的导语中进行匹配,获取匹配成功导语作为当前第二样本标题的新的导语;将所述第二样本标题作为输入,与所述第二样本标题对应的导语以及所述新的导语作为输出,基于预设算法训练出第一生成模型。
- 一种基于深度学习的目标软文的生成装置,其特征在于,所述装置包括:标题匹配模块,用于接收目标对象的相关信息,根据所述相关信息从标题库中匹配出若干条适配的目标标题,所述标题库中的标题由采集到的标题通过第三生成模型扩展而来;导语生成模块,用于将所述目标标题输入到第一生成模型中,生成至少一个目标导语;正文生成模块,用于根据所述相关信息以及预设规则生成至少一个符合预设结构的输入信息,将所述输入信息输入到第二生成模型中,生成至少一个目标正文;信息组装模块,对所述目标标题、所述目标导语以及所述目标正文进行组装,获取多篇目标软文。
- 根据权利要求6所述的基于深度学习的目标软文的生成装置,其特征在于,所述正文生成模块包括:第一分词单元,用于对所述相关信息进行分词处理,从获取到的第一分词结果中提取出满足预设条件的目标分词;分词重组单元,用于对所述目标分词进行重组,获取至少一个符合预设结构的输入信息;正文生成单元,用于将所述输入信息输入到第二生成模型中,生成至少一个目标正文。
- 根据权利要求6或7所述的基于深度学习的目标软文的生成装置,其特征在于,所述装置还包括第一构建模块,包括:第二分词单元,用于对采集到的若干第一样本标题进行分词处理,获取第二分词结果;第一提取单元,用于采用预设的第一关键词提取方法从所述第一样本标题中提取出第一关键词;标题生成单元,用于将所述第二分词结果以及所述第一关键词输入到第三生成模型,获取多个新的标题,所述标题库由所述新的标题构成。
- 根据权利要求8所述的基于深度学习的目标软文的生成装置,其特征在于,所述第一构建模块还包括:第一求交单元,用于对所述第一关键词集合与所述第二分词结果取交集,获取输入数据 集合;第一训练单元,用于将所述输入数据集合的数据作为输入,所述目标标题作为输出,基于预设算法训练出第三生成模型。
- 根据权利要求6或7所述的基于深度学习的目标软文的生成装置,其特征在于,所述装置还包括第二构建模块,包括:第三分词单元,用于对采集到的若干第二样本标题以及与所述第二样本标题对应的导语对进行分词处理;第二提取单元,用于采用预设的第二关键词提取方法从所述第二样本标题中提取出第二关键词;第二求交单元,用于对所述第二关键词集合与每条分词后的所述第二样本标题取交集,获取目标关键词;导语拓展单元,用于遍历每一所述第二样本标题,将所述目标关键词与全量分词后的与所述第二样本标题对应的导语中进行匹配,获取匹配成功导语作为当前第二样本标题的新的导语;第二训练单元,用于将所述第二样本标题作为输入,与所述第二样本标题对应的导语以及所述新的导语作为输出,基于预设算法训练出第一生成模型。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CA3166556A CA3166556A1 (en) | 2019-12-30 | 2020-06-19 | Method and device for generating target advertorial based on deep learning |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911403246.2A CN111178018B (zh) | 2019-12-30 | 2019-12-30 | 一种基于深度学习的目标软文的生成方法及装置 |
CN201911403246.2 | 2019-12-30 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021135091A1 true WO2021135091A1 (zh) | 2021-07-08 |
Family
ID=70650585
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/097007 WO2021135091A1 (zh) | 2019-12-30 | 2020-06-19 | 一种基于深度学习的目标软文的生成方法及装置 |
Country Status (3)
Country | Link |
---|---|
CN (1) | CN111178018B (zh) |
CA (1) | CA3166556A1 (zh) |
WO (1) | WO2021135091A1 (zh) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111178018B (zh) * | 2019-12-30 | 2024-03-26 | 苏宁云计算有限公司 | 一种基于深度学习的目标软文的生成方法及装置 |
CN115409000B (zh) * | 2022-11-02 | 2023-01-24 | 浪潮通信信息系统有限公司 | 一种热点人物软文自动生成方法及装置 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103246710A (zh) * | 2013-04-22 | 2013-08-14 | 张经纶 | 一种多媒体旅游游记的自动生成方法及装置 |
CN106777193A (zh) * | 2016-12-23 | 2017-05-31 | 李鹏 | 一种自动撰写特定稿件的方法 |
CN109992764A (zh) * | 2017-12-29 | 2019-07-09 | 阿里巴巴集团控股有限公司 | 一种文案生成方法及装置 |
US20190236148A1 (en) * | 2018-02-01 | 2019-08-01 | Jungle Disk, L.L.C. | Generative text using a personality model |
CN110162623A (zh) * | 2019-04-15 | 2019-08-23 | 深圳壹账通智能科技有限公司 | 软文自动生成方法、装置、计算机设备及存储介质 |
CN111178018A (zh) * | 2019-12-30 | 2020-05-19 | 苏宁云计算有限公司 | 一种基于深度学习的目标软文的生成方法及装置 |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106503255B (zh) * | 2016-11-15 | 2020-05-12 | 科大讯飞股份有限公司 | 基于描述文本自动生成文章的方法及系统 |
CN109388745A (zh) * | 2018-06-15 | 2019-02-26 | 云天弈(北京)信息技术有限公司 | 一种批量文章自动写作系统 |
CN109460447A (zh) * | 2018-11-29 | 2019-03-12 | 上海文军信息技术有限公司 | 一种营销软文识别方法 |
-
2019
- 2019-12-30 CN CN201911403246.2A patent/CN111178018B/zh active Active
-
2020
- 2020-06-19 WO PCT/CN2020/097007 patent/WO2021135091A1/zh active Application Filing
- 2020-06-19 CA CA3166556A patent/CA3166556A1/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103246710A (zh) * | 2013-04-22 | 2013-08-14 | 张经纶 | 一种多媒体旅游游记的自动生成方法及装置 |
CN106777193A (zh) * | 2016-12-23 | 2017-05-31 | 李鹏 | 一种自动撰写特定稿件的方法 |
CN109992764A (zh) * | 2017-12-29 | 2019-07-09 | 阿里巴巴集团控股有限公司 | 一种文案生成方法及装置 |
US20190236148A1 (en) * | 2018-02-01 | 2019-08-01 | Jungle Disk, L.L.C. | Generative text using a personality model |
CN110162623A (zh) * | 2019-04-15 | 2019-08-23 | 深圳壹账通智能科技有限公司 | 软文自动生成方法、装置、计算机设备及存储介质 |
CN111178018A (zh) * | 2019-12-30 | 2020-05-19 | 苏宁云计算有限公司 | 一种基于深度学习的目标软文的生成方法及装置 |
Also Published As
Publication number | Publication date |
---|---|
CN111178018B (zh) | 2024-03-26 |
CA3166556A1 (en) | 2021-07-08 |
CN111178018A (zh) | 2020-05-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109726293B (zh) | 一种因果事件图谱构建方法、系统、装置及存储介质 | |
CN108984683B (zh) | 结构化数据的提取方法、系统、设备及存储介质 | |
CN110765759B (zh) | 意图识别方法及装置 | |
CN108959418A (zh) | 一种人物关系抽取方法、装置、计算机装置及计算机可读存储介质 | |
CN108363725B (zh) | 一种用户评论观点提取和观点标签生成的方法 | |
CN107273358B (zh) | 一种基于管道模式的端到端英文篇章结构自动分析方法 | |
CN111709242B (zh) | 一种基于命名实体识别的中文标点符号添加方法 | |
CN101458681A (zh) | 语音翻译方法和语音翻译装置 | |
WO2018153215A1 (zh) | 一种自动生成语义相近句子样本的方法 | |
WO2021135091A1 (zh) | 一种基于深度学习的目标软文的生成方法及装置 | |
CN103116578A (zh) | 一种融合句法树和统计机器翻译技术的翻译方法与装置 | |
WO2022226716A1 (zh) | 基于深度学习的Java程序内部注释的生成方法及系统 | |
Amancio et al. | An analysis of crowdsourced text simplifications | |
CN106980620A (zh) | 一种对中文字串进行匹配的方法及装置 | |
Wei et al. | Poet-based poetry generation: Controlling personal style with recurrent neural networks | |
CN108519963B (zh) | 一种将流程模型自动转换为多语言文本的方法 | |
CN116432654A (zh) | 一种基于内容上下文的自动续写生成方法 | |
CN111914555A (zh) | 基于Transformer结构的自动化关系抽取系统 | |
CN113343717A (zh) | 一种基于翻译记忆库的神经机器翻译方法 | |
CN113688606A (zh) | 一种自动化进行文档报告写作的方法 | |
Asscher | The explanatory power of descriptive translation studies in the machine translation era | |
CN111708896B (zh) | 一种应用于生物医学文献的实体关系抽取方法 | |
CN113963306A (zh) | 基于人工智能的课件片头制作方法和装置 | |
CN107491443B (zh) | 一种包含非常规词汇的中文句子翻译方法及系统 | |
CN112784579A (zh) | 一种基于数据增强的阅读理解选择题答题方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20910378 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 3166556 Country of ref document: CA |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20910378 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205 DATED 23/08/2022) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20910378 Country of ref document: EP Kind code of ref document: A1 |