WO2024007476A1 - Design concept generation network construction method and automatic concept scheme generation method - Google Patents

Design concept generation network construction method and automatic concept scheme generation method Download PDF

Info

Publication number
WO2024007476A1
WO2024007476A1 PCT/CN2022/125347 CN2022125347W WO2024007476A1 WO 2024007476 A1 WO2024007476 A1 WO 2024007476A1 CN 2022125347 W CN2022125347 W CN 2022125347W WO 2024007476 A1 WO2024007476 A1 WO 2024007476A1
Authority
WO
WIPO (PCT)
Prior art keywords
importance
design
sequence
decoder
input vocabulary
Prior art date
Application number
PCT/CN2022/125347
Other languages
French (fr)
Chinese (zh)
Inventor
赵武
于淼
郭鑫
张凯
赵乾
於辉
李骏
赖兵
姜冲
蒋亦伟
伍博
陈星宇
Original Assignee
四川大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 四川大学 filed Critical 四川大学
Priority to US18/120,434 priority Critical patent/US20240005130A1/en
Publication of WO2024007476A1 publication Critical patent/WO2024007476A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F30/00Computer-aided design [CAD]
    • G06F30/20Design optimisation, verification or simulation
    • G06F30/27Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y04INFORMATION OR COMMUNICATION TECHNOLOGIES HAVING AN IMPACT ON OTHER TECHNOLOGY AREAS
    • Y04SSYSTEMS INTEGRATING TECHNOLOGIES RELATED TO POWER NETWORK OPERATION, COMMUNICATION OR INFORMATION TECHNOLOGIES FOR IMPROVING THE ELECTRICAL POWER GENERATION, TRANSMISSION, DISTRIBUTION, MANAGEMENT OR USAGE, i.e. SMART GRIDS
    • Y04S10/00Systems supporting electrical power generation, transmission or distribution
    • Y04S10/50Systems or methods supporting the power network operation or management, involving a certain degree of interaction with the load-side end user applications

Definitions

  • the invention belongs to the technical field of product design and relates to the automatic generation of concept solutions in product design, and in particular to the construction of a design concept generation network and the automatic generation of concept solutions based on the generation network.
  • Prior design data is an important source of innovation.
  • concept solution generation is the process of absorbing valuable design knowledge from prior design data, and further migrating and reorganizing cross-domain design knowledge to generate creative concept solutions.
  • the engineering data used in conceptual design is increasing day by day, bringing a rich source of innovation to the research on conceptual scheme generation.
  • Fully applying it to the conceptual scheme generation stage will be beneficial to Expand the design space and generate more design concepts. But it also faces more severe challenges, mainly in two aspects: on the one hand, with the explosive growth of design data, the amount of knowledge applied to conceptual design is also gradually increasing.
  • design knowledge mainly comes from the description of existing product design solutions in different fields, which often presents complex Diversity, such as rich knowledge types such as functions, structures, scientific effects, cases, etc., and the relationships between knowledge are also more complex and flexible. How to screen out valuable design knowledge based on design problems or design constraints and combine multiple types of cross-domain design knowledge to generate new conceptual solutions is becoming increasingly difficult.
  • the purpose of the present invention is to provide a method of constructing a design concept generation network and an automatic generation method of concept solutions based on the generation network, which can generate concepts from massive texts based on design problems. Adaptively learn potential rules such as reasoning, transfer, and reorganization of cross-domain design knowledge in the data, and automatically generate conceptual solutions in text form, thereby reducing reliance on designers' manual experience and improving design efficiency.
  • the invention provides a Design Concept Generation Network (DCGN), further performs network training and learning, and finally automatically generates concept solutions by inputting design problems into the trained DCGN network.
  • DCGN Design Concept Generation Network
  • the present invention adopts the following technical solutions.
  • the design concept generation network construction method provided by the present invention is based on the self-attention mechanism of the Transformer network and cleverly introduces vocabulary importance constraints to construct a new generation network;
  • the design concept generation network includes a Transformer encoder, Transformer decoder, importance constraint matrix generation module, importance constraint embedding layer, cross-attention layer and optimization module;
  • the present invention uses training sample set data to train the design concept generation network;
  • the training sample set data includes several Samples, each sample includes input vocabulary and target sequence;
  • the design concept generation network construction method includes the following steps:
  • S1 uses the Transformer encoder to obtain the hidden layer features of the encoder based on the input vocabulary in the sample;
  • S2 uses the Transformer decoder to obtain the decoder hidden layer features based on the target sequence in the sample;
  • S3 uses the importance constraint matrix generation module to obtain the importance constraint matrix based on the input vocabulary and target sequence in the sample;
  • S4 uses the importance constraint embedding layer to map the importance constraint matrix to the distributed vector space to obtain two input vocabulary importance embedding features
  • S5 uses the cross-attention layer to obtain the generated sequence based on the encoder hidden layer features, decoder hidden layer features and two input vocabulary importance embedding features;
  • S6 constructs a loss function based on the generation sequence and the target sequence, and uses the optimization module to adjust the network parameters based on the loss function; then repeat steps S1-S6 until the loss function meets the set requirements, and the design concept generation network is obtained.
  • the Transformer encoder converts discrete input words into words through the self-attention layer. (m represents the number of input words in the current sample, n represents the dimension of the input word embedding vector) mapped to the distributed feature representation, that is, the encoder hidden layer features are obtained (d represents the number of neurons in the hidden layer. In the present invention, the number of neurons in the hidden layer of the Transformer encoder and Transformer decoder is designed to be the same):
  • SA() represents spatial attention; Represent the weight matrix of the Transformer encoder's self-attention layer respectively.
  • x is discrete and unordered, so there is no need to incorporate the position embedding in the graph when calculating he , and the output he will not contain any position information.
  • M represents the maximum number of input words contained in the samples in the entire training sample set.
  • SA() represents spatial attention; Respectively represent the weight matrix of the Transformer encoder's self-attention layer; y :t-1 represents the target sequence at time (t-1) during the training process.
  • the importance constraint matrix in the present invention is represented by C, which is the input vocabulary information and the target sequence at different times.
  • the result of joint action can be expressed as:
  • y 0 is the sequence given at the initial moment, which can be generated using special characters such as ⁇ EOS>;
  • y :t represents the target sequence before time t (including time t) in the sample;
  • T represents the length of the target sequence in the sample.
  • represents the vector or matrix dot product operation; is the relative importance vector of input x in the target sequence y :t , which can be calculated as:
  • the relative importance value after the above regularization process is an integer.
  • the present invention introduces two new importance constraint embedding matrices. and Map the importance constraint matrix C constructed above to the distributed vector space to obtain two input vocabulary importance embedding features. and Therefore, at the t-th moment of generation, there is:
  • equations (7) and (8) are based on the relative importance Index importance constraint matrix and The corresponding row, the default row is set to zero, and the feature is obtained
  • step S5 across the attention layer (Cross-Attention layer, CA), the encoder hidden layer features (h e ) and the decoder hidden layer features are fused and two input vocabulary importance embedding features (preferably two input vocabulary importance embedding features in the present invention ), get the generated sequence at the current time t
  • the j-th element in the CA function can be expressed as:
  • step S6 for the given N samples
  • the loss function of DCGN constructed based on the generated sequence and the target sequence is:
  • the network parameters here mainly refer to the weight matrix of the encoder self-attention layer used to obtain the encoder hidden layer features, the weight matrix and the importance constraint embedding matrix of the decoder self-attention layer used to obtain the decoder hidden layer features. .
  • the initialization parameters of the importance constraint embedding matrix can be implemented through random initialization.
  • the initialization parameters of the weight matrix of the encoder self-attention layer used to obtain the encoder hidden layer features and the weight matrix of the decoder self-attention layer used to obtain the decoder hidden layer features can be implemented by random initialization; in the preferred implementation , the weight matrix of the encoder self-attention layer and the weight matrix of the decoder self-attention layer are obtained by using the common sense text database to perform regular Transformer networks (such as T5 (Text-to-Text Transfer Transformer), GPT (Generative Pre-trained Transformer), etc.), so that the design concept generation network provided by the present invention has the ability to understand common sense knowledge and ensure the fluency of the design concept description generated by the design concept generation network DCGN.
  • the method provided by the present invention is then used to further design the concept generation network DCGN for training, which can enable the network model to have intelligent reasoning capabilities of engineering design knowledge and ensure the rationality of the generated design concept description.
  • the present invention further provides a method for automatically generating concept solutions, using the constructed design concept generation network to perform operations in accordance with the following steps:
  • L1 based on the input vocabulary, use the Transformer encoder to obtain the hidden layer features of the encoder;
  • L4 use the importance constraint embedding layer to map the importance constraint matrix to the distributed vector space, and obtain two input vocabulary importance embedding features
  • L5 based on the encoder hidden layer features, decoder hidden layer features and two input vocabulary importance embedding features, use cross-attention layers to obtain the generated sequence.
  • the input vocabulary can be composed of keywords from the construction of design problems, or it can be composed of more than one design incentive, or it can be composed of keywords from the construction of design requirements, or at least a combination of the above two sources of input vocabulary.
  • step L2 based on the sequence generated at the previous moment, the hidden layer features of the decoder at the current moment are calculated according to the following formula
  • y :t-1 represents the input sequence of the decoder at time t during the generation process, Represents the sequence given at the initial moment.
  • Special characters such as ⁇ EOS> can be used to represent the generation. Represents the sequence generated at the previous moment.
  • the constraint matrix is calculated in time steps based on the actual sequence generated at each moment.
  • the importance constraint matrix C :t-1 is calculated according to the following formula:
  • x represents the input vocabulary
  • y :t-1 represents the input sequence of the decoder at time t during the generation process
  • the order of vocabulary importance is given in advance, and can also be set to be consistent.
  • step L4 two input vocabulary importance embedding features are used Calculate the importance embedding features of the two input words at the current moment according to the above formulas (7) and (8).
  • step L5 the generated sequence at the current time is calculated according to formulas (9)-(10).
  • the present invention has the following beneficial effects:
  • the vocabulary importance constraint matrix proposed by this invention records the input vocabulary constraint information contained in the generated text sequence, which can effectively ensure the reliability and effectiveness of the generated concept solution;
  • the present invention proposes an importance constraint embedding layer, which maps the constructed importance constraint matrix to a distributed vector space, and uses continuous real number vectors to represent the relative importance of input words in the generated sequence or target sequence, which is beneficial to capturing potential importance information at the semantic level to realize knowledge reasoning at the semantic level;
  • the cross-attention layer constructed by the present invention maps the input vocabulary importance embedding features to the generated sequence to supervise the generation of text sequences containing input vocabulary importance information.
  • Figure 1 is a schematic diagram of the construction and use framework of the design concept generation network in the embodiment of the present invention.
  • Figure 2 is a schematic diagram of the principle of a method for constructing a design concept generation network in an embodiment of the present invention.
  • Figure 3 is a schematic diagram of the concept method generation process in the embodiment of the present invention.
  • this embodiment first uses web crawler technology to obtain text data and perform preprocessing; then constructs a DCGN model to further train and learn the model; and finally inputs the design into the trained DCGN model. Keywords are used as input vocabulary to automatically generate concept plans.
  • This embodiment first uses conventional web crawler technology to crawl massive text data from websites, such as scientific papers, patents, etc., and selects sentences of a certain length from the collected text data as the corpus for this study. Then the text data is preprocessed, and a keyword extraction algorithm is used to obtain a certain number of keywords (excluding stop words) and their importance from each sentence. Finally, each sentence and its corresponding keyword information are combined into a sample pair, and an entire sample set composed of sample pairs is constructed for subsequent network training. Each sample uses the extracted keywords as the input sequence and the corresponding sentence as the target sequence.
  • the design concept generation network includes a Transformer encoder, a Transformer decoder, and an importance constraint matrix generation module, importance constraint embedding layer, cross-attention layer and optimization module.
  • the Transformer encoder is used to obtain the encoder hidden layer features;
  • the Transformer decoder is used to obtain the decoder hidden layer features;
  • the importance constraint matrix generation module is used to generate the importance constraint matrix;
  • the importance constraint embedding layer is used to map the importance constraint matrix to Distributed vector space is used to obtain two input word importance embedding features;
  • the cross-attention layer is used to obtain the generated sequence;
  • the optimization module is used to optimize network parameters based on the loss function.
  • the design concept generation network construction method provided in this embodiment mainly involves training and learning on the training sample set to obtain the weight matrix of the encoder self-attention layer for obtaining the characteristics of the encoder hidden layer, and the weight matrix for obtaining the decoder hidden layer.
  • the decoder of layer features has a weight matrix from the attention layer and two importance constraint embedding matrices.
  • This embodiment uses the common sense text database (selected from Wikipedia) to train the conventional Transformer network (T5 (Text-to-Text Transfer Transformer)) to obtain the weight matrix of the encoder self-attention layer used to obtain the encoder hidden layer features. and the weight matrix initialization parameters of the decoder self-attention layer used to obtain the decoder hidden layer features.
  • T5 Text-to-Text Transfer Transformer
  • the two importance constraint embedding matrices obtain their initialization parameters through random initialization.
  • the common sense text database is used to train the T5 network to obtain the weight matrix of the encoder's self-attention layer. and the weight matrix of the decoder self-attention layer Explanation of encoders and decoders as mentioned earlier.
  • T5 network training please refer to the literature Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer (Colin Raffel et al, Journal of Machine Learning Research 21 (2020) 1-67. Take the trained T5 network The weight matrix of the encoder self-attention layer in and the weight matrix of the decoder self-attention layer As the design concept of the present invention, corresponding initialization parameters of the network are generated.
  • the design concept generation network construction method provided by this embodiment includes the following steps:
  • the importance constraint matrix C is determined by the formula (3) given previously.
  • the importance constraint matrix at time t-1 is obtained based on the input vocabulary and the target sequence y :t at time t-1, that is, It can be calculated according to the previous formulas (4)-(6).
  • the importance embedding features of the two input words at time t are calculated according to formulas (7) and (8).
  • the loss function of DCGN is calculated according to formula (11). Based on this loss function, adjust and optimize the network parameters through the conventional Adam optimization algorithm, and then repeat steps S1-S6 until the loss function meets the set requirements. For example, the loss function tends to be stable and basically unchanged, and the design concept generation is completed. Network construction.
  • the DCGN network model After the DCGN network model is fully trained, it has the ability to express knowledge and reason, and can adaptively absorb, transfer, and reorganize cross-domain design knowledge. At this stage, well-defined design problems or valuable knowledge incentives are input into the trained DCGN, and relevant design concept descriptions can be automatically generated.
  • the DCGN network model combines design knowledge from different fields to generate design concepts that not only contain design input information, but also ensure the novelty and creativity of the generated design concepts.
  • L1 based on the input vocabulary, uses the Transformer encoder to obtain the hidden layer features of the encoder.
  • the encoder hidden layer feature he is calculated according to the above formula (1).
  • L2 based on the sequence generated at the previous moment, uses the Transformer decoder to obtain the decoder output hidden layer features at the current moment.
  • the importance constraint matrix C :t-1 is calculated according to formula (13).
  • the absolute importance of the input words in the decoder input sequence y :t-1 is set to be consistent, and the value of w i is 1.
  • L4 use the importance constraint embedding layer to map the importance constraint matrix to the distributed vector space, and obtain two input vocabulary importance embedding features.
  • L5 based on the encoder hidden layer features, decoder hidden layer features and two input vocabulary importance embedding features, use cross-attention layers to obtain the generated sequence.
  • the generated sequence at the current moment is calculated according to formulas (9)-(10).
  • the specific concept plan generation stage is that the output vocabulary at the previous moment will be used as a new part of the input at the current moment, and new vocabulary will be generated in sequence until the end identifier ⁇ EOS> is encountered.
  • the process is shown in Figure 3.
  • Use x ⁇ drone,deliver,life,preserver ⁇ as the input vocabulary, and use special characters such as ⁇ EOS> to represent the initial moment generation sequence
  • C is calculated in time steps based on the actual sequence generated at each moment, regardless of the target sequence, which is completely different from the training phase.
  • the design problem of this case is to provide edible water sources for residents in coastal areas.
  • 10 graduate students majoring in mechanical engineering were invited to define the design problem using a limited number of keywords.
  • the design team unanimously agreed to use "purification” (purification or purify), “desalination” (desalination or desalinate), “solar” (solar), “seawater” (seawater) and
  • the "drink” keyword is used to define the design problem.
  • the corresponding design concept can be automatically generated by using the constructed DCGN network according to the above-mentioned concept scheme automatic generation method. The results are shown in Table 2.
  • the automatically generated design concepts provide more specific and feasible design concepts, such as inventing a purification system for purifying seawater into drinking water, or using solar energy to desalinate seawater for the production of canned drinking water or beverage products.
  • design concepts provide residents or businesses in coastal areas with early design ideas for product development.
  • design problems involved in the present invention can also be composed of design incentives.
  • design incentives provide rich and valuable design inspiration.
  • the transition from design incentives to conceptual solutions often relies on the designer's rich experience and knowledge, and generates Conceptual solutions are inefficient and the process becomes very difficult for inexperienced novice designers.
  • Some design incentives for UAVs obtained in this embodiment are shown in Table 3.
  • the automatically generated conceptual scheme is shown in Table 4. Since there are many types of combinations, only some valuable conceptual solutions are shown and analyzed here. For example:
  • the DCGN network automatically generates the design concept "a drone rescue radar system is disclosed that is capable of detecting the presence of an animal in the “nearby of the drone using bio”.
  • the DCGN network automatically generates the design concept "the drone may also be configured to receive ground fire data from the ground drone and to determine a location of the fire in response to detecting the resulting fire.”, and “the drone may also be configured to receive ground fire data from the ground drone and to determine a location of the fire in response to determining the terrain”.
  • design requirements are critical in determining the design direction of a new product.
  • Online product review data provides accurate, reliable, authentic information for analyzing design needs, and is easily accessible.
  • the keywords “disinfection” (sterilization), “temperature” (temperature), “function” (function), and “capacity” (capacity) are used here as the design issues of the DCGN network model, automatically
  • the generated conceptual scheme is shown in Table 6. It is easy to find that using different input keyword combinations will generate different conceptual solutions. More importantly, all automatically generated conceptual solutions contain the entered design problem keywords, and some feasible and creative conceptual solutions have been generated, such as using ion exchangers to improve sterilization and disinfection capabilities. , which meets the design needs to a certain extent.
  • Table 5 The top 30 most frequent demand keywords in online user review data
  • DCGN Design Concept Generation Network

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Computer Hardware Design (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Geometry (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Machine Translation (AREA)
  • Document Processing Apparatus (AREA)

Abstract

Disclosed in the present invention are a design concept generation network construction method and an automatic concept scheme generation method. A design concept generation network comprises a Transformer encoder, a Transformer decoder, an importance constraint matrix generation module, an importance constraint embedding layer, a cross-attention layer, and an optimization module. According to the present invention, on the basis of an attention mechanism of Transformer, a vocabulary importance constraint is ingeniously introduced, input vocabulary constraint information comprised in a generated text sequence is recorded, and the reliability and effectiveness of a generated concept scheme can be effectively guaranteed; and potential importance information at a semantic level can be captured, and knowledge reasoning at the semantic level can be realized.

Description

设计概念生成网络构建方法及概念方案自动生成方法Design concept generation network construction method and concept scheme automatic generation method 技术领域Technical field
本发明属于产品设计技术领域,涉及产品设计中的概念方案自动生成,尤其涉及构建一种设计概念生成网络,并基于该生成网络自动生成概念方案。The invention belongs to the technical field of product design and relates to the automatic generation of concept solutions in product design, and in particular to the construction of a design concept generation network and the automatic generation of concept solutions based on the generation network.
背景技术Background technique
在先设计数据是重要的创新源泉。作为产品创新概念设计的核心,概念方案生成就是从在先设计数据中吸收有价值的设计知识,并进一步迁移、重组跨领域设计知识以产生创造性概念方案的过程。随着大数据、大知识时代的到来,应用于概念设计的工程数据正在日益增长,给概念方案生成的研究带也来了丰富的创新源泉,充分将其应用于概念方案生成阶段,将有利于扩大设计空间,产生更多的设计概念。但也面对更严峻的挑战,主要表现在两个方面:一方面,随着设计数据的爆炸式增长,应用于概念设计的知识量也在逐渐增加,以设计师的人工经验和设计启发式为主对大量的设计知识进行推理、迁移和重组以产生创造性的概念方案变得越来越困难;另一方面,设计知识主要来源于不同领域的现有产品设计方案的描述,往往呈现出复杂多样性,如功能、结构、科学效应、案例等丰富的知识类型,并且知识间关联关系也是更加复杂、灵活的。如何根据设计问题或设计约束筛选出有价值的设计知识,并通过组合多类型的跨领域设计知识以产生新的概念方案变得越来越难。Prior design data is an important source of innovation. As the core of product innovation conceptual design, concept solution generation is the process of absorbing valuable design knowledge from prior design data, and further migrating and reorganizing cross-domain design knowledge to generate creative concept solutions. With the advent of the era of big data and big knowledge, the engineering data used in conceptual design is increasing day by day, bringing a rich source of innovation to the research on conceptual scheme generation. Fully applying it to the conceptual scheme generation stage will be beneficial to Expand the design space and generate more design concepts. But it also faces more severe challenges, mainly in two aspects: on the one hand, with the explosive growth of design data, the amount of knowledge applied to conceptual design is also gradually increasing. Based on the designer’s manual experience and design heuristics, It has become increasingly difficult to reason, transfer and reorganize a large amount of design knowledge to produce creative conceptual solutions; on the other hand, design knowledge mainly comes from the description of existing product design solutions in different fields, which often presents complex Diversity, such as rich knowledge types such as functions, structures, scientific effects, cases, etc., and the relationships between knowledge are also more complex and flexible. How to screen out valuable design knowledge based on design problems or design constraints and combine multiple types of cross-domain design knowledge to generate new conceptual solutions is becoming increasingly difficult.
随着深度学习技术的快速发展,许多自动生成技术被开发并成功完成各种智能任务,如:机器翻译、图像生成、语音识别等。最新的深度生成模型也已经在工程设计领域的多个方面取得重要突破,如结构优化、材料设计、形状合成等。也存在一些研究使用拓扑优化和生成对抗网络等生成模型来自动生成图像和空间形状等形式的设计概念。这些设计概念要么过于抽象,难以理解;要么过于详细,不适合早期阶段的概念方案设计探索。With the rapid development of deep learning technology, many automatic generation technologies have been developed and successfully completed various intelligent tasks, such as machine translation, image generation, speech recognition, etc. The latest deep generative models have also achieved important breakthroughs in many aspects of engineering design, such as structural optimization, material design, shape synthesis, etc. There are also studies using generative models such as topology optimization and generative adversarial networks to automatically generate design concepts in the form of images and spatial shapes. These design concepts are either too abstract to be understood or too detailed to be suitable for early stage conceptual design exploration.
研究发现,文本是最通用的、最常见的设计概念描述形式,能够覆盖丰富的、有价值的设计知识。如何通过简单、有效的模型从海量跨领域文本数据中学习设计知识间潜在的推理、迁移和重组等组合规律,生成适用于早期阶段的概念方案,是目前产品设计需要丞待解决的一个重要问题。Research has found that text is the most versatile and common form of design concept description, and can cover rich and valuable design knowledge. How to learn the potential combination rules of reasoning, migration and reorganization between design knowledge from massive cross-domain text data through simple and effective models to generate conceptual solutions suitable for the early stages is an important issue that needs to be solved in current product design. .
发明内容Contents of the invention
针对目前产品设计领域缺少概念方案自动生成方法的技术现状,本发明的目的旨在提供一种设计概念生成网络构建方法及基于该生成网络的概念方案自动生成方法,能够根据设计问题,从海量文本数据中自适应学习跨领域设计知识的推理、迁移、重组等潜在的规律,自动生成文本形式的概念方案,从而减少对设计师人工经验的依赖,提高设计效率。In view of the current technical status quo of the lack of automatic generation method of concept solutions in the field of product design, the purpose of the present invention is to provide a method of constructing a design concept generation network and an automatic generation method of concept solutions based on the generation network, which can generate concepts from massive texts based on design problems. Adaptively learn potential rules such as reasoning, transfer, and reorganization of cross-domain design knowledge in the data, and automatically generate conceptual solutions in text form, thereby reducing reliance on designers' manual experience and improving design efficiency.
本发明的发明思路为:本发明提供一种设计概念生成网络(Design Concept Generation Network,DCGN),进一步进行网络的训练和学习,最后通过向训练完成的DCGN网络中输入设计问题自动生成概念方案。The idea of the invention is as follows: the invention provides a Design Concept Generation Network (DCGN), further performs network training and learning, and finally automatically generates concept solutions by inputting design problems into the trained DCGN network.
为达到上述目的,本发明采用以下技术方案来实现。In order to achieve the above objects, the present invention adopts the following technical solutions.
本发明提供的设计概念生成网络构建方法,是在Transformer网络的自注意力机制基础上,巧妙地引入了词汇重要度约束,构建一种新的生成网络;该设计概念生成网络包括Transformer编码器,Transformer解码器,重要度约束矩阵生成模块,重要度约束嵌入层,跨注意力层和优化模块;本发明使用训练用样本集数据对设计概念生成网络进行训练;所述训练用样本集数据包括若干样本,每个样本包括输入词汇和目标序列;所述设计概念生成网络构建方法包括以下步骤:The design concept generation network construction method provided by the present invention is based on the self-attention mechanism of the Transformer network and cleverly introduces vocabulary importance constraints to construct a new generation network; the design concept generation network includes a Transformer encoder, Transformer decoder, importance constraint matrix generation module, importance constraint embedding layer, cross-attention layer and optimization module; the present invention uses training sample set data to train the design concept generation network; the training sample set data includes several Samples, each sample includes input vocabulary and target sequence; the design concept generation network construction method includes the following steps:
S1依据样本中输入词汇,利用Transformer编码器得到编码器隐藏层特征;S1 uses the Transformer encoder to obtain the hidden layer features of the encoder based on the input vocabulary in the sample;
S2依据样本中目标序列,利用Transformer解码器得到解码器隐藏层特征;S2 uses the Transformer decoder to obtain the decoder hidden layer features based on the target sequence in the sample;
S3依据样本中输入词汇和目标序列,利用重要度约束矩阵生成模块获取重要度约束矩阵;S3 uses the importance constraint matrix generation module to obtain the importance constraint matrix based on the input vocabulary and target sequence in the sample;
S4利用重要度约束嵌入层将重要度约束矩阵映射至分布式向量空间,得到两个输入词汇重要 度嵌入特征;S4 uses the importance constraint embedding layer to map the importance constraint matrix to the distributed vector space to obtain two input vocabulary importance embedding features;
S5依据编码器隐藏层特征、解码器隐藏层特征和两个输入词汇重要度嵌入特征,利用跨注意力层获取生成序列;S5 uses the cross-attention layer to obtain the generated sequence based on the encoder hidden layer features, decoder hidden layer features and two input vocabulary importance embedding features;
S6依据生成序列和目标序列构建损失函数,并基于损失函数利用优化模块对网络参数进行调整;然后重复步骤S1-S6,直至损失函数满足设定要求,得到设计概念生成网络。S6 constructs a loss function based on the generation sequence and the target sequence, and uses the optimization module to adjust the network parameters based on the loss function; then repeat steps S1-S6 until the loss function meets the set requirements, and the design concept generation network is obtained.
上述步骤S1中,Transformer编码器通过自注意力层将离散的输入词汇
Figure PCTCN2022125347-appb-000001
(m表示当前样本中输入词汇的数目,n表示输入词汇嵌入向量的维度)映射至分布式的特征表示,即得到编码器隐藏层特征
Figure PCTCN2022125347-appb-000002
(d表示隐藏层的神经元个数,本发明中Transformer编码器和Transformer解码器隐藏层的神经元个数设计为相同):
In the above step S1, the Transformer encoder converts discrete input words into words through the self-attention layer.
Figure PCTCN2022125347-appb-000001
(m represents the number of input words in the current sample, n represents the dimension of the input word embedding vector) mapped to the distributed feature representation, that is, the encoder hidden layer features are obtained
Figure PCTCN2022125347-appb-000002
(d represents the number of neurons in the hidden layer. In the present invention, the number of neurons in the hidden layer of the Transformer encoder and Transformer decoder is designed to be the same):
Figure PCTCN2022125347-appb-000003
Figure PCTCN2022125347-appb-000003
式中,SA()表示空间注意力;
Figure PCTCN2022125347-appb-000004
分别表示Transformer编码器自注意力层的权重矩阵。x是离散的、无序的,因此在计算h e时不需要并入图中的位置嵌入,输出的h e将不包含任何位置信息。当计算得到的h e向量维度m小于M时,使用0向量补齐,使得
Figure PCTCN2022125347-appb-000005
M≥m>1,M表示整个训练用样本集中样本包含的最多输入词汇数目。
In the formula, SA() represents spatial attention;
Figure PCTCN2022125347-appb-000004
Represent the weight matrix of the Transformer encoder's self-attention layer respectively. x is discrete and unordered, so there is no need to incorporate the position embedding in the graph when calculating he , and the output he will not contain any position information. When the calculated he vector dimension m is less than M, use 0 vectors to complete it, so that
Figure PCTCN2022125347-appb-000005
M≥m>1, M represents the maximum number of input words contained in the samples in the entire training sample set.
上述步骤S2中,Transformer解码器通过自注意力层将前一时刻的目标序列y :t-1=[y 0,y 1,L,y t-1]映射至分布式的特征表示,即得到解码器隐藏层特征
Figure PCTCN2022125347-appb-000006
In the above step S2, the Transformer decoder maps the target sequence y :t-1 = [y 0 , y 1 , L, y t-1 ] at the previous moment to the distributed feature representation through the self-attention layer, that is, we get Decoder hidden layer features
Figure PCTCN2022125347-appb-000006
Figure PCTCN2022125347-appb-000007
Figure PCTCN2022125347-appb-000007
式中,SA()表示空间注意力;
Figure PCTCN2022125347-appb-000008
分别表示Transformer编码器自注意力层的权重矩阵;y :t-1表示训练过程中(t-1)时刻的目标序列。
In the formula, SA() represents spatial attention;
Figure PCTCN2022125347-appb-000008
Respectively represent the weight matrix of the Transformer encoder's self-attention layer; y :t-1 represents the target sequence at time (t-1) during the training process.
上述公式(1)和(2)中的SA()函数可通过如下公式计算:The SA() function in the above formulas (1) and (2) can be calculated by the following formula:
Figure PCTCN2022125347-appb-000009
Figure PCTCN2022125347-appb-000009
对于编码器,K表示
Figure PCTCN2022125347-appb-000010
V表示
Figure PCTCN2022125347-appb-000011
Q表示
Figure PCTCN2022125347-appb-000012
对于解码器,K表示
Figure PCTCN2022125347-appb-000013
V表示
Figure PCTCN2022125347-appb-000014
Q表示
Figure PCTCN2022125347-appb-000015
For the encoder, K means
Figure PCTCN2022125347-appb-000010
V means
Figure PCTCN2022125347-appb-000011
Q means
Figure PCTCN2022125347-appb-000012
For the decoder, K means
Figure PCTCN2022125347-appb-000013
V means
Figure PCTCN2022125347-appb-000014
Q means
Figure PCTCN2022125347-appb-000015
上述步骤S3中,本发明中重要度约束矩阵用C表示,其是输入词汇信息与不同时刻的目标序列
Figure PCTCN2022125347-appb-000016
共同作用的结果,可以表示为:
In the above step S3, the importance constraint matrix in the present invention is represented by C, which is the input vocabulary information and the target sequence at different times.
Figure PCTCN2022125347-appb-000016
The result of joint action can be expressed as:
Figure PCTCN2022125347-appb-000017
Figure PCTCN2022125347-appb-000017
式中,y 0为初始时刻给定的序列,可以用特殊字符例如<EOS>来表示生成;
Figure PCTCN2022125347-appb-000018
表示目 标序列y :t中包含的输入词汇重要度约束向量,即C :t;y :t表示样本中t时刻以前(包含t时刻)的目标序列;T表示样本中目标序列长度。
In the formula, y 0 is the sequence given at the initial moment, which can be generated using special characters such as <EOS>;
Figure PCTCN2022125347-appb-000018
Represents the input vocabulary importance constraint vector contained in the target sequence y :t , that is, C :t ; y :t represents the target sequence before time t (including time t) in the sample; T represents the length of the target sequence in the sample.
Figure PCTCN2022125347-appb-000019
可计算为:
Figure PCTCN2022125347-appb-000019
Can be calculated as:
Figure PCTCN2022125347-appb-000020
Figure PCTCN2022125347-appb-000020
式中,·表示向量或矩阵点积运算;
Figure PCTCN2022125347-appb-000021
为输入x在目标序列y :t中相对重要度向量,可计算为:
In the formula, · represents the vector or matrix dot product operation;
Figure PCTCN2022125347-appb-000021
is the relative importance vector of input x in the target sequence y :t , which can be calculated as:
Figure PCTCN2022125347-appb-000022
Figure PCTCN2022125347-appb-000022
式中,
Figure PCTCN2022125347-appb-000023
表示第i个输入词汇在目标序列y :t中相对重要度;w i表示第i个输入词汇在目标序列y :t中绝对重要度;w min表示输入词汇在目标序列y :t中的最小绝对重要度;w max表示输入词汇在目标序列y :t中的最大绝对重要度;[]为取整运算。
In the formula,
Figure PCTCN2022125347-appb-000023
represents the relative importance of the i-th input vocabulary in the target sequence y :t ; w i represents the absolute importance of the i-th input vocabulary in the target sequence y: t ; w min represents the minimum value of the input vocabulary in the target sequence y :t Absolute importance; w max represents the maximum absolute importance of the input vocabulary in the target sequence y :t ; [] is the rounding operation.
经过上述正则化处理后的相对重要度值
Figure PCTCN2022125347-appb-000024
为整数。
The relative importance value after the above regularization process
Figure PCTCN2022125347-appb-000024
is an integer.
另外,
Figure PCTCN2022125347-appb-000025
表示目标序列y :t中包含的输入词汇约束;当目标序列y :t中包含了输入词汇的第i个词时,则向量c t中的第i个元素为1,可以通过如下方式计算:
in addition,
Figure PCTCN2022125347-appb-000025
Represents the input vocabulary constraints contained in the target sequence y :t ; when the target sequence y :t contains the i-th word of the input vocabulary, the i-th element in the vector c t is 1, which can be calculated as follows:
Figure PCTCN2022125347-appb-000026
Figure PCTCN2022125347-appb-000026
所以根据式(3)计算得到的
Figure PCTCN2022125347-appb-000027
为相对重要度组成的整数向量。
Therefore, calculated according to formula (3)
Figure PCTCN2022125347-appb-000027
is an integer vector composed of relative importance.
上述步骤S4中,本发明中引入两个新的重要度约束嵌入矩阵
Figure PCTCN2022125347-appb-000028
Figure PCTCN2022125347-appb-000029
将上述构建的重要度约束矩阵C映射至分布式向量空间,得到两个输入词汇重要度嵌入特征
Figure PCTCN2022125347-appb-000030
Figure PCTCN2022125347-appb-000031
所以,在生成的第t个时刻,有:
In the above step S4, the present invention introduces two new importance constraint embedding matrices.
Figure PCTCN2022125347-appb-000028
and
Figure PCTCN2022125347-appb-000029
Map the importance constraint matrix C constructed above to the distributed vector space to obtain two input vocabulary importance embedding features.
Figure PCTCN2022125347-appb-000030
and
Figure PCTCN2022125347-appb-000031
Therefore, at the t-th moment of generation, there is:
Figure PCTCN2022125347-appb-000032
Figure PCTCN2022125347-appb-000032
Figure PCTCN2022125347-appb-000033
Figure PCTCN2022125347-appb-000033
式中,t∈{1,2,…,T}。另外,式(7)、(8)根据相对重要度
Figure PCTCN2022125347-appb-000034
索引重要度约束矩阵
Figure PCTCN2022125347-appb-000035
Figure PCTCN2022125347-appb-000036
相应的行,缺省行置零,得到特征
Figure PCTCN2022125347-appb-000037
In the formula, t∈{1,2,…,T}. In addition, equations (7) and (8) are based on the relative importance
Figure PCTCN2022125347-appb-000034
Index importance constraint matrix
Figure PCTCN2022125347-appb-000035
and
Figure PCTCN2022125347-appb-000036
The corresponding row, the default row is set to zero, and the feature is obtained
Figure PCTCN2022125347-appb-000037
上述步骤S5中,跨注意力层(Cross-Attention layer,CA),融合处理编码器隐藏层特征(h e)、解码器隐藏层特征
Figure PCTCN2022125347-appb-000038
和两个输入词汇重要度嵌入特征(本发明中优选为两个输入词汇重要度嵌 入特征
Figure PCTCN2022125347-appb-000039
),得到当前时刻t的生成序列
Figure PCTCN2022125347-appb-000040
In the above step S5, across the attention layer (Cross-Attention layer, CA), the encoder hidden layer features (h e ) and the decoder hidden layer features are fused
Figure PCTCN2022125347-appb-000038
and two input vocabulary importance embedding features (preferably two input vocabulary importance embedding features in the present invention
Figure PCTCN2022125347-appb-000039
), get the generated sequence at the current time t
Figure PCTCN2022125347-appb-000040
Figure PCTCN2022125347-appb-000041
Figure PCTCN2022125347-appb-000041
式中,
Figure PCTCN2022125347-appb-000042
表示解码器自注意力层的权重矩阵。
In the formula,
Figure PCTCN2022125347-appb-000042
Represents the weight matrix of the decoder’s self-attention layer.
在具体实现方式中,CA函数中第j个元素可表示为:In the specific implementation, the j-th element in the CA function can be expressed as:
Figure PCTCN2022125347-appb-000043
Figure PCTCN2022125347-appb-000043
式中,
Figure PCTCN2022125347-appb-000044
Figure PCTCN2022125347-appb-000045
In the formula,
Figure PCTCN2022125347-appb-000044
Figure PCTCN2022125347-appb-000045
随着时间的推移,重复上述步骤S2-S5,当t=T时,DCGN得到了最终文本生成序列
Figure PCTCN2022125347-appb-000046
As time goes by, the above steps S2-S5 are repeated. When t=T, DCGN obtains the final text generation sequence.
Figure PCTCN2022125347-appb-000046
对训练用样本集中的样本,重复上述步骤S1-S5,便可得到不同样本对应的生成序列。For the samples in the training sample set, repeat the above steps S1-S5 to obtain the generated sequences corresponding to different samples.
步骤S6中,对于给定的N个样本
Figure PCTCN2022125347-appb-000047
依据生成序列和目标序列构建的DCGN的损失函数为:
In step S6, for the given N samples
Figure PCTCN2022125347-appb-000047
The loss function of DCGN constructed based on the generated sequence and the target sequence is:
Figure PCTCN2022125347-appb-000048
Figure PCTCN2022125347-appb-000048
式中,
Figure PCTCN2022125347-appb-000049
表示在t时刻的生成序列
Figure PCTCN2022125347-appb-000050
和对应时刻的目标序列y :t之间误差,通常采用交叉熵计算。
In the formula,
Figure PCTCN2022125347-appb-000049
Represents the generated sequence at time t
Figure PCTCN2022125347-appb-000050
The error between it and the target sequence y :t at the corresponding time is usually calculated by cross entropy.
基于上述损失函数,通过Adam优化算法对网络参数进行调整优化,之后重复步骤S1-S6,直至损失函数满足设定要求,例如损失函数趋于稳定、基本不变,便完成了对设计概念生成网络的构建。这里的网络参数主要是指用于获取编码器隐藏层特征的编码器自注意力层的权重矩阵、用于获取解码器隐藏层特征的解码器自注意力层的权重矩阵和重要度约束嵌入矩阵。重要度约束嵌入矩阵的初始化参数可以通过随机初始化实现。用于获取编码器隐藏层特征的编码器自注意力层的权重矩阵和用于获取解码器隐藏层特征的解码器自注意力层的权重矩阵的初始化参数可以通过随机初始化实现;在优选实现方式中,编码器自注意力层的权重矩阵和解码器自注意力层的权重矩阵是通过利用常识文本数据库对常规Transformer网络(例如T5(Text-to-Text Transfer Transformer)、GPT(Generative Pre-trained Transformer)等)进行训练得到,这样可以使本发明提供的设计概念生成网络具有理解常识知识的能力,确保设计概念生成网络DCGN生成的设计概念描述的流畅性。再利用本发明提供的方法进一步设计概念生成网络DCGN进行训练,能够使该网络模型具有工程设计知识的智能推理能力,确保生成的设计概念描述的合理性。Based on the above loss function, adjust and optimize the network parameters through the Adam optimization algorithm, and then repeat steps S1-S6 until the loss function meets the set requirements. For example, the loss function tends to be stable and basically unchanged, and the design concept generation network is completed. of construction. The network parameters here mainly refer to the weight matrix of the encoder self-attention layer used to obtain the encoder hidden layer features, the weight matrix and the importance constraint embedding matrix of the decoder self-attention layer used to obtain the decoder hidden layer features. . The initialization parameters of the importance constraint embedding matrix can be implemented through random initialization. The initialization parameters of the weight matrix of the encoder self-attention layer used to obtain the encoder hidden layer features and the weight matrix of the decoder self-attention layer used to obtain the decoder hidden layer features can be implemented by random initialization; in the preferred implementation , the weight matrix of the encoder self-attention layer and the weight matrix of the decoder self-attention layer are obtained by using the common sense text database to perform regular Transformer networks (such as T5 (Text-to-Text Transfer Transformer), GPT (Generative Pre-trained Transformer), etc.), so that the design concept generation network provided by the present invention has the ability to understand common sense knowledge and ensure the fluency of the design concept description generated by the design concept generation network DCGN. The method provided by the present invention is then used to further design the concept generation network DCGN for training, which can enable the network model to have intelligent reasoning capabilities of engineering design knowledge and ensure the rationality of the generated design concept description.
本发明进一步提供了一种概念方案自动生成方法,使用构建的设计概念生成网络按照以下步骤执行操作:The present invention further provides a method for automatically generating concept solutions, using the constructed design concept generation network to perform operations in accordance with the following steps:
L1,依据输入词汇,利用Transformer编码器得到编码器隐藏层特征;L1, based on the input vocabulary, use the Transformer encoder to obtain the hidden layer features of the encoder;
L2,依据上一时刻生成序列,利用Transformer解码器得到当前时刻解码器隐藏层特征;L2, based on the sequence generated at the previous moment, use the Transformer decoder to obtain the hidden layer features of the decoder at the current moment;
L3,依据样本中输入词汇和上一时刻生成序列,利用重要度约束矩阵生成模块获取重要度约束矩阵;L3, based on the input vocabulary in the sample and the generated sequence at the previous moment, use the importance constraint matrix generation module to obtain the importance constraint matrix;
L4,利用重要度约束嵌入层将重要度约束矩阵映射至分布式向量空间,得到两个输入词汇重要度嵌入特征;L4, use the importance constraint embedding layer to map the importance constraint matrix to the distributed vector space, and obtain two input vocabulary importance embedding features;
L5,依据编码器隐藏层特征、解码器隐藏层特征和两个输入词汇重要度嵌入特征,利用跨注意力层获取生成序列。L5, based on the encoder hidden layer features, decoder hidden layer features and two input vocabulary importance embedding features, use cross-attention layers to obtain the generated sequence.
上述步骤L1中,输入词汇可以由来自于设计问题构建的关键词组成,也可以由一个以上设计激励组成,也可以来自于设计需求构建的关键词组成,或者至少上述两种输入词汇来源的组合。In the above step L1, the input vocabulary can be composed of keywords from the construction of design problems, or it can be composed of more than one design incentive, or it can be composed of keywords from the construction of design requirements, or at least a combination of the above two sources of input vocabulary. .
上述步骤L2中,依据上一时刻生成序列,按照以下公式计算得到当前时刻解码器隐藏层特征
Figure PCTCN2022125347-appb-000051
In the above step L2, based on the sequence generated at the previous moment, the hidden layer features of the decoder at the current moment are calculated according to the following formula
Figure PCTCN2022125347-appb-000051
Figure PCTCN2022125347-appb-000052
Figure PCTCN2022125347-appb-000052
式中,y :t-1表示生成过程中t时刻解码器的输入序列,
Figure PCTCN2022125347-appb-000053
表示初始时刻给定的序列,可以用特殊字符例如<EOS>来表示生成,
Figure PCTCN2022125347-appb-000054
表示前一时刻生成序列。
In the formula, y :t-1 represents the input sequence of the decoder at time t during the generation process,
Figure PCTCN2022125347-appb-000053
Represents the sequence given at the initial moment. Special characters such as <EOS> can be used to represent the generation.
Figure PCTCN2022125347-appb-000054
Represents the sequence generated at the previous moment.
上述步骤L3中,概念方案生成阶段,约束矩阵是根据每一时刻生成的实际序列按照时间步骤计算得到的。In the above step L3, in the conceptual solution generation stage, the constraint matrix is calculated in time steps based on the actual sequence generated at each moment.
依据样本中输入词汇和上一时刻生成序列,按照以下公式计算得到重要度约束矩阵C :t-1Based on the input vocabulary in the sample and the sequence generated at the previous moment, the importance constraint matrix C :t-1 is calculated according to the following formula:
Figure PCTCN2022125347-appb-000055
Figure PCTCN2022125347-appb-000055
式中,x表示输入词汇,y :t-1表示生成过程中t时刻解码器的输入序列,
Figure PCTCN2022125347-appb-000056
为输入x在解码器输入序列y :t-1中相对重要度向量,可以按照上述公式(5)计算得到,其中输入词汇在解码器输入序列y :t-1中的绝对重要度可以按照输入词汇重要性顺序事先给定,也可以均设置为一致。
In the formula, x represents the input vocabulary, y :t-1 represents the input sequence of the decoder at time t during the generation process,
Figure PCTCN2022125347-appb-000056
is the relative importance vector of input x in the decoder input sequence y :t-1 , which can be calculated according to the above formula (5), where the absolute importance of the input vocabulary in the decoder input sequence y :t-1 can be calculated according to the input The order of vocabulary importance is given in advance, and can also be set to be consistent.
上述步骤L4中,使用的是两个输入词汇重要度嵌入特征
Figure PCTCN2022125347-appb-000057
按照上述公式(7)和(8)计算得到当前时刻两个输入词汇重要度嵌入特征
Figure PCTCN2022125347-appb-000058
In the above step L4, two input vocabulary importance embedding features are used
Figure PCTCN2022125347-appb-000057
Calculate the importance embedding features of the two input words at the current moment according to the above formulas (7) and (8).
Figure PCTCN2022125347-appb-000058
上述步骤L5中,按照公式(9)-(10)计算得到当前时刻生成序列。In the above step L5, the generated sequence at the current time is calculated according to formulas (9)-(10).
重复上述步骤L1-L5,至生成序列长度满足设定要求或者遇到结束标识符<EOS>,得到最终生成序列,即概念方案。Repeat the above steps L1-L5 until the length of the generated sequence meets the set requirements or the end identifier <EOS> is encountered, and the final generated sequence, that is, the conceptual solution is obtained.
与现有技术相比,本发明具有以下有益效果:Compared with the prior art, the present invention has the following beneficial effects:
1)本发明在Transformer的注意力机制基础上,巧妙地引入了词汇重要度约束,构建了一种新的设计概念生成网络,1) Based on the attention mechanism of Transformer, this invention cleverly introduces vocabulary importance constraints and constructs a new design concept generation network.
2)本发明提出的词汇重要度约束矩阵,记录了生成的文本序列中包含的输入词汇约束信息,能够有效保证生成概念方案的可靠性和有效性;2) The vocabulary importance constraint matrix proposed by this invention records the input vocabulary constraint information contained in the generated text sequence, which can effectively ensure the reliability and effectiveness of the generated concept solution;
3)本发明提出了重要度约束嵌入层,将构建的重要度约束矩阵映射至分布式向量空间,使用连续的实数向量表示输入词汇在生成序列或目标序列中的相对重要性,有利于捕获潜在的、语义层面的重要度信息,实现语义层面的知识推理;3) The present invention proposes an importance constraint embedding layer, which maps the constructed importance constraint matrix to a distributed vector space, and uses continuous real number vectors to represent the relative importance of input words in the generated sequence or target sequence, which is beneficial to capturing potential importance information at the semantic level to realize knowledge reasoning at the semantic level;
4)本发明构建的跨注意力层,将输入词汇重要度嵌入特征映射至生成的序列中,以监督生成包含输入词汇重要度信息的文本序列。4) The cross-attention layer constructed by the present invention maps the input vocabulary importance embedding features to the generated sequence to supervise the generation of text sequences containing input vocabulary importance information.
附图说明Description of the drawings
图1本发明实施例中设计概念生成网络构建及使用框架示意图。Figure 1 is a schematic diagram of the construction and use framework of the design concept generation network in the embodiment of the present invention.
图2为本发明实施例中设计概念生成网络构建方法原理示意图。Figure 2 is a schematic diagram of the principle of a method for constructing a design concept generation network in an embodiment of the present invention.
图3为本发明实施例中概念方法生成过程示意图。Figure 3 is a schematic diagram of the concept method generation process in the embodiment of the present invention.
具体实施方式Detailed ways
结合附图对本发明各实施例的技术方案进行清楚、完整的描述,显然,所描述实施例仅仅是本发明的一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动的前提下所得到的所有其它实施例,都属于本发明。The technical solutions of various embodiments of the present invention are clearly and completely described with reference to the accompanying drawings. Obviously, the described embodiments are only some of the embodiments of the present invention, rather than all of the embodiments. Based on the embodiments of the present invention, all other embodiments obtained by those of ordinary skill in the art without any creative efforts belong to the present invention.
实施例1Example 1
如图1所示,本实施例首先是使用网络爬虫技术进行获取文本数据并进行预处理;然后进行DCGN模型的构建,进一步进行模型的训练和学习;最后通过向训练完成的DCGN模型中输入设计关键词作为输入词汇,自动生成概念方案。As shown in Figure 1, this embodiment first uses web crawler technology to obtain text data and perform preprocessing; then constructs a DCGN model to further train and learn the model; and finally inputs the design into the trained DCGN model. Keywords are used as input vocabulary to automatically generate concept plans.
本实施例首先利用常规网络爬虫技术从网站上爬取海量的文本数据,比如说科技论文、专利等,并从收集的文本数据中筛选出一定长度的句子作为本研究的语料库。然后进行文本数据预处理,进而使用关键词提取算法从每个句子中获取一定数目的关键词(排除停用词)及其重要度。最后将每一个句子与其对应的关键词信息组成样本对,构建样本对组成的全体样本集,用于后续的网络训练,每个样本以提取的关键词作为输入序列,对应的句子作为目标序列。This embodiment first uses conventional web crawler technology to crawl massive text data from websites, such as scientific papers, patents, etc., and selects sentences of a certain length from the collected text data as the corpus for this study. Then the text data is preprocessed, and a keyword extraction algorithm is used to obtain a certain number of keywords (excluding stop words) and their importance from each sentence. Finally, each sentence and its corresponding keyword information are combined into a sample pair, and an entire sample set composed of sample pairs is constructed for subsequent network training. Each sample uses the extracted keywords as the input sequence and the corresponding sentence as the target sequence.
(一)设计概念生成网络的构建(1) Construction of design concept generation network
本实施例在Transformer网络的自注意力机制基础上,巧妙地引入了词汇重要度约束,构建一种新的生成网络;该设计概念生成网络包括Transformer编码器,Transformer解码器,重要度约束矩阵生成模块,重要度约束嵌入层,跨注意力层和优化模块。Transformer编码器用于获取编码器隐藏层特征;Transformer解码器用于获取解码器隐藏层特征;重要度约束矩阵生成模块用于生成重要度约束矩阵;重要度约束嵌入层用于将重要度约束矩阵映射至分布式向量空间,得到两个输入词汇重要度嵌入特征;跨注意力层用于获取生成序列;优化模块用于依据损失函数对网络参数进行优化。This embodiment cleverly introduces vocabulary importance constraints based on the self-attention mechanism of the Transformer network to construct a new generation network; the design concept generation network includes a Transformer encoder, a Transformer decoder, and an importance constraint matrix generation module, importance constraint embedding layer, cross-attention layer and optimization module. The Transformer encoder is used to obtain the encoder hidden layer features; the Transformer decoder is used to obtain the decoder hidden layer features; the importance constraint matrix generation module is used to generate the importance constraint matrix; the importance constraint embedding layer is used to map the importance constraint matrix to Distributed vector space is used to obtain two input word importance embedding features; the cross-attention layer is used to obtain the generated sequence; the optimization module is used to optimize network parameters based on the loss function.
本实施例提供的设计概念生成网络构建方法,主要是通过对训练用样本集进行训练学习,得到用于获取编码器隐藏层特征的编码器自注意力层的权重矩阵、用于获取解码器隐藏层特征的解码器自注意力层的权重矩阵和两个重要度约束嵌入矩阵。The design concept generation network construction method provided in this embodiment mainly involves training and learning on the training sample set to obtain the weight matrix of the encoder self-attention layer for obtaining the characteristics of the encoder hidden layer, and the weight matrix for obtaining the decoder hidden layer. The decoder of layer features has a weight matrix from the attention layer and two importance constraint embedding matrices.
本实施例利用常识文本数据库(选自维基百科)对常规Transformer网络(T5(Text-to-Text Transfer Transformer))进行训练得到用于获取编码器隐藏层特征的编码器自注意力层的权重矩阵和用于获取解码器隐藏层特征的解码器自注意力层的权重矩阵初始化参数。两个重要度约束嵌入矩阵通过随机初始化得到其初始化参数。This embodiment uses the common sense text database (selected from Wikipedia) to train the conventional Transformer network (T5 (Text-to-Text Transfer Transformer)) to obtain the weight matrix of the encoder self-attention layer used to obtain the encoder hidden layer features. and the weight matrix initialization parameters of the decoder self-attention layer used to obtain the decoder hidden layer features. The two importance constraint embedding matrices obtain their initialization parameters through random initialization.
1、使用常识文本数据库对T5网络进行训练1. Use the common sense text database to train the T5 network
这里利用常识文本数据库对T5网络进行训练,得到编码器自注意力层的权重矩阵
Figure PCTCN2022125347-appb-000059
Figure PCTCN2022125347-appb-000060
和解码器自注意力层的权重矩阵
Figure PCTCN2022125347-appb-000061
关于编码器和解码器的解释,如前所述。对T5网络训练的具体实现过程参见文献Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer(Colin Raffel et al,Journal of Machine Learning Research 21(2020)1-67。以训练后的T5网络中的编码器自注意力层的权重矩阵
Figure PCTCN2022125347-appb-000062
和解码器自注意力层的权重矩阵
Figure PCTCN2022125347-appb-000063
作为本发明设计概念生成网络相应初始化参数。
Here, the common sense text database is used to train the T5 network to obtain the weight matrix of the encoder's self-attention layer.
Figure PCTCN2022125347-appb-000059
Figure PCTCN2022125347-appb-000060
and the weight matrix of the decoder self-attention layer
Figure PCTCN2022125347-appb-000061
Explanation of encoders and decoders as mentioned earlier. For the specific implementation process of T5 network training, please refer to the literature Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer (Colin Raffel et al, Journal of Machine Learning Research 21 (2020) 1-67. Take the trained T5 network The weight matrix of the encoder self-attention layer in
Figure PCTCN2022125347-appb-000062
and the weight matrix of the decoder self-attention layer
Figure PCTCN2022125347-appb-000063
As the design concept of the present invention, corresponding initialization parameters of the network are generated.
2、设计概念生成网络构建2. Design concept generation network construction
如图2所示,本实施例提供的设计概念生成网络构建方法包括以下步骤:As shown in Figure 2, the design concept generation network construction method provided by this embodiment includes the following steps:
S1,依据样本中输入词汇,利用Transformer编码器得到编码器隐藏层特征。S1, based on the input vocabulary in the sample, use the Transformer encoder to obtain the hidden layer features of the encoder.
本步骤依据样本中的输入词汇x={x 1,x 2,L,x m},按照前面公式(1)计算得到编码器隐藏层特征h eThis step is based on the input vocabulary x={x 1 , x 2 , L, x m } in the sample and calculates the encoder hidden layer feature he according to the previous formula (1).
S2,依据样本中目标序列,利用Transformer解码器得到解码器隐藏层特征。S2, based on the target sequence in the sample, use the Transformer decoder to obtain the decoder hidden layer features.
本步骤依据样本中的目标序列y :t-1=[y 0,y 1,L,y t-1],按照前面公式(2)计算得到t时刻解码器隐藏层特征
Figure PCTCN2022125347-appb-000064
This step is based on the target sequence y in the sample : t-1 = [y 0 , y 1 , L, y t-1 ], and calculates the hidden layer features of the decoder at time t according to the previous formula (2)
Figure PCTCN2022125347-appb-000064
S3,依据样本中输入词汇和目标序列,利用重要度约束矩阵生成模块获取重要度约束矩阵。S3, based on the input vocabulary and target sequence in the sample, use the importance constraint matrix generation module to obtain the importance constraint matrix.
重要度约束矩阵C由前面给出的公式(3)确定。The importance constraint matrix C is determined by the formula (3) given previously.
这里是依据输入词汇和t-1时刻的目标序列y :t得到t-1时刻的重要度约束矩阵,即
Figure PCTCN2022125347-appb-000065
可以按照前面公式(4)-(6)计算得到。
Here, the importance constraint matrix at time t-1 is obtained based on the input vocabulary and the target sequence y :t at time t-1, that is,
Figure PCTCN2022125347-appb-000065
It can be calculated according to the previous formulas (4)-(6).
下面用一个具体实例展示在DCGN训练过程中C的详细计算过程。假设DCGN的输入为三个关键词构成的集合{“sensor”,“device”,“sowing”},生成的目标序列为“a sensor device for determing a position of seeds while sowing.”。假设输入词汇在目标序列中的重要度为w=[0.9,0.7,0.5],且M=5,则根据公式(4)可计算相对重要度向量为
Figure PCTCN2022125347-appb-000066
该值代表着这三个输入词汇在目标序列的相对重要度。接下来说明C的计算步骤,如表1所示:
The following uses a specific example to show the detailed calculation process of C during DCGN training. Assume that the input of DCGN is a set of three keywords {"sensor", "device", "sowing"}, and the generated target sequence is "a sensor device for determining a position of seeds while sowing." Assuming that the importance of the input vocabulary in the target sequence is w=[0.9,0.7,0.5], and M=5, the relative importance vector can be calculated according to formula (4) as
Figure PCTCN2022125347-appb-000066
This value represents the relative importance of these three input words in the target sequence. Next, the calculation steps of C are explained, as shown in Table 1:
a)在开始生成起始符<EOS>时,该时刻所对应的目标序列中不包含任何输入词汇,所以此时c 0为全零向量。由式(4)计算的
Figure PCTCN2022125347-appb-000067
也为全零向量,对应表1中第一列数值;
a) When starting to generate the start symbol <EOS>, the target sequence corresponding to this moment does not contain any input vocabulary, so c 0 is an all-zero vector at this time. Calculated by formula (4)
Figure PCTCN2022125347-appb-000067
It is also an all-zero vector, corresponding to the first column value in Table 1;
b)因为第二个生成的目标词为“a”,该时刻的目标序列中仍不包含任何输入词汇,,所以此时c 1为全零向量。由式(4)计算的
Figure PCTCN2022125347-appb-000068
也为全零向量,对应表1中第二列数值;
b) Because the second generated target word is "a", the target sequence at this moment still does not contain any input words, so c 1 is an all-zero vector at this time. Calculated by formula (4)
Figure PCTCN2022125347-appb-000068
It is also an all-zero vector, corresponding to the second column value in Table 1;
c)因为第三个生成的目标词为“sensor”,该时刻的目标序列中仅包含输入词汇中的“sensor”,所以有c 2=[1;0;0],再由公式(4)可计算
Figure PCTCN2022125347-appb-000069
对应表1中第三列数值;
c) Because the third generated target word is "sensor", the target sequence at this moment only contains "sensor" in the input vocabulary, so c 2 = [1; 0; 0], and then according to formula (4) Computable
Figure PCTCN2022125347-appb-000069
Corresponds to the third column value in Table 1;
d)因为第四个生成的目标词为“device”,该时刻的目标序列中包含输入词汇中的“sensor”和“device”,所以有c 3=[1;1;0],再由公式(4)可计算
Figure PCTCN2022125347-appb-000070
对应表1中第四列数值;
d) Because the fourth generated target word is "device", the target sequence at this moment contains "sensor" and "device" in the input vocabulary, so there is c 3 = [1; 1; 0], and then according to the formula (4) Calculable
Figure PCTCN2022125347-appb-000070
Corresponds to the fourth column value in Table 1;
e)依次类推,直到产生结束符<EOS>。e) And so on until the end character <EOS> is generated.
表1 DCGN网络构建过程中C的生成过程Table 1 The generation process of C during the construction of DCGN network
Figure PCTCN2022125347-appb-000071
Figure PCTCN2022125347-appb-000071
S4,利用重要度约束嵌入层将重要度约束矩阵映射至分布式向量空间,得到两个输入词汇重要度嵌入特征。S4, use the importance constraint embedding layer to map the importance constraint matrix to the distributed vector space, and obtain two input vocabulary importance embedding features.
本步骤中,按照公式(7)和(8)计算得到t时刻两个输入词汇重要度嵌入特征
Figure PCTCN2022125347-appb-000072
Figure PCTCN2022125347-appb-000073
In this step, the importance embedding features of the two input words at time t are calculated according to formulas (7) and (8).
Figure PCTCN2022125347-appb-000072
and
Figure PCTCN2022125347-appb-000073
S5,依据编码器隐藏层特征、解码器隐藏层特征和两个输入词汇重要度嵌入特征,利用跨注意力层获取生成序列。S5, based on the encoder hidden layer features, the decoder hidden layer features and the two input vocabulary importance embedding features, use the cross-attention layer to obtain the generated sequence.
本步骤中,按照公式(9)和(10)计算得到t时刻的生成序列
Figure PCTCN2022125347-appb-000074
In this step, the generated sequence at time t is calculated according to formulas (9) and (10)
Figure PCTCN2022125347-appb-000074
随着时间的推移,重复上述步骤S2-S5,当t=T时,DCGN得到了最终文本生成序列
Figure PCTCN2022125347-appb-000075
As time goes by, the above steps S2-S5 are repeated. When t=T, DCGN obtains the final text generation sequence.
Figure PCTCN2022125347-appb-000075
对训练用样本集给定的N个样本
Figure PCTCN2022125347-appb-000076
重复上述步骤S1-S5,便可得到N个样本对应的生成序列。
For N samples given in the training sample set
Figure PCTCN2022125347-appb-000076
Repeat the above steps S1-S5 to obtain the generated sequences corresponding to N samples.
S6,依据生成序列和目标序列构建损失函数,并基于损失函数对网络参数进行调整;然后重复步骤S1-S6,直至损失函数满足设定要求,得到设计概念生成网络。S6: Construct a loss function based on the generation sequence and the target sequence, and adjust the network parameters based on the loss function; then repeat steps S1-S6 until the loss function meets the set requirements, and the design concept generation network is obtained.
本步骤中,对于给定的N个样本,按照公式(11)计算得到DCGN的损失函数。基于该损失函数,通过常规Adam优化算法对网络参数进行调整优化,之后重复步骤S1-S6,直至损失函数满足设定要求,例如损失函数趋于稳定、基本不变,便完成了对设计概念生成网络的构建。In this step, for the given N samples, the loss function of DCGN is calculated according to formula (11). Based on this loss function, adjust and optimize the network parameters through the conventional Adam optimization algorithm, and then repeat steps S1-S6 until the loss function meets the set requirements. For example, the loss function tends to be stable and basically unchanged, and the design concept generation is completed. Network construction.
DCGN网络模型经过充分训练之后,就具备了知识表达与推理的能力,能够自适应地吸收、迁移、重组跨领域设计知识。在该阶段,将定义良好的设计问题或有价值的知识激励等输入至训练完成的DCGN中,就能够自动生成相关的设计概念描述。DCGN网络模型通过组合不同领域的设计知识,生成的设计概念中不仅包含设计输入信息,而且确保了所生成的设计概念的新颖性和创造性。After the DCGN network model is fully trained, it has the ability to express knowledge and reason, and can adaptively absorb, transfer, and reorganize cross-domain design knowledge. At this stage, well-defined design problems or valuable knowledge incentives are input into the trained DCGN, and relevant design concept descriptions can be automatically generated. The DCGN network model combines design knowledge from different fields to generate design concepts that not only contain design input information, but also ensure the novelty and creativity of the generated design concepts.
(二)设计概念生成网络的测试。(2) Testing of the design concept generation network.
下面通过输入设计问题(即关键词)来检验所提出的概念方案自动生成方法的有效性和实用性。Next, the validity and practicability of the proposed automatic generation method of conceptual solutions are tested by inputting design questions (i.e., keywords).
本实施例提供的概念方案自动生成方法,使用构建的设计概念生成网络按照以下步骤执行操作:The method for automatically generating concept solutions provided in this embodiment uses the constructed design concept generation network to perform operations according to the following steps:
L1,依据输入词汇,利用Transformer编码器得到编码器隐藏层特征。L1, based on the input vocabulary, uses the Transformer encoder to obtain the hidden layer features of the encoder.
本步骤中,按照上述公式(1)计算得到编码器隐藏层特征h eIn this step, the encoder hidden layer feature he is calculated according to the above formula (1).
L2,依据上一时刻生成序列,利用Transformer解码器得到当前时刻解码器输隐藏层特征。L2, based on the sequence generated at the previous moment, uses the Transformer decoder to obtain the decoder output hidden layer features at the current moment.
本步骤中,按照上述公式(12)计算得到解码器隐藏层特征
Figure PCTCN2022125347-appb-000077
In this step, the decoder hidden layer features are calculated according to the above formula (12)
Figure PCTCN2022125347-appb-000077
L3,依据样本中输入词汇和上一时刻生成序列,利用重要度约束矩阵生成模块获取重要度约束矩阵。L3, based on the input vocabulary in the sample and the generated sequence at the previous moment, use the importance constraint matrix generation module to obtain the importance constraint matrix.
本步骤中,按照公式(13)计算得到重要度约束矩阵C :t-1In this step, the importance constraint matrix C :t-1 is calculated according to formula (13).
本实施例中,输入词汇在解码器输入序列y :t-1中的绝对重要度均设置为一致,w i取值为1。 In this embodiment, the absolute importance of the input words in the decoder input sequence y :t-1 is set to be consistent, and the value of w i is 1.
L4,利用重要度约束嵌入层将重要度约束矩阵映射至分布式向量空间,得到两个输入词汇重要度嵌入特征。L4, use the importance constraint embedding layer to map the importance constraint matrix to the distributed vector space, and obtain two input vocabulary importance embedding features.
本步骤中,使用的是两个输入词汇重要度嵌入特征
Figure PCTCN2022125347-appb-000078
按照上述公式(7)和(8)计算得到当前时刻两个输入词汇重要度嵌入特征
Figure PCTCN2022125347-appb-000079
In this step, two input vocabulary importance embedding features are used
Figure PCTCN2022125347-appb-000078
Calculate the importance embedding features of the two input words at the current moment according to the above formulas (7) and (8).
Figure PCTCN2022125347-appb-000079
L5,依据编码器隐藏层特征、解码器隐藏层特征和两个输入词汇重要度嵌入特征,利用跨注意力层获取生成序列。L5, based on the encoder hidden layer features, decoder hidden layer features and two input vocabulary importance embedding features, use cross-attention layers to obtain the generated sequence.
本步骤中,按照公式(9)-(10)计算得到当前时刻生成序列。In this step, the generated sequence at the current moment is calculated according to formulas (9)-(10).
重复上述步骤L1-L5,至生成序列长度满足设定要求或者遇到结束标识符<EOS>,得到最终生成序列,即概念方案。Repeat the above steps L1-L5 until the length of the generated sequence meets the set requirements or the end identifier <EOS> is encountered, and the final generated sequence, that is, the conceptual solution is obtained.
因此,具体的概念方案生成阶段是上一时刻的输出词汇将作为当前时刻输入的新增部分,依次生成新的词汇,直到遇到结束标识符<EOS>,其过程如图3所示。以x={drone,deliver,life,preserver}作为输入词汇,以特殊字符例<EOS>来表示初始时刻生成序列
Figure PCTCN2022125347-appb-000080
重复上述步骤L1-L5,直至遇到结束标识符<EOS>,得到生成序列
Figure PCTCN2022125347-appb-000081
Therefore, the specific concept plan generation stage is that the output vocabulary at the previous moment will be used as a new part of the input at the current moment, and new vocabulary will be generated in sequence until the end identifier <EOS> is encountered. The process is shown in Figure 3. Use x={drone,deliver,life,preserver} as the input vocabulary, and use special characters such as <EOS> to represent the initial moment generation sequence
Figure PCTCN2022125347-appb-000080
Repeat the above steps L1-L5 until the end identifier <EOS> is encountered and the generated sequence is obtained
Figure PCTCN2022125347-appb-000081
在生成阶段C是根据每一时刻生成的实际序列按照时间步计算的,跟目标序列无关,这与训练阶段是完全不同的。In the generation phase, C is calculated in time steps based on the actual sequence generated at each moment, regardless of the target sequence, which is completely different from the training phase.
下面列举不同输入词汇来源生成概念方案的具体实例:The following are specific examples of concept generation schemes from different input vocabulary sources:
1、本案例的设计问题是为沿海地区的居民提供可食用的水源。为了更准确、简洁地表达设计问题,邀请了10位机械工程专业的研究生同学,使用有限个关键词定义该设计问题。考虑到沿海地区日照、光线充足的优势,设计团队一致同意使用“净化”(purification或purify)、“海水淡化”(desalination或desalinate)、“太阳的”(solar)、“海水”(seawater)和“饮用”(drink)关键词来定义该设计问题。通过组合不同的关键词作为设计输入,利用构建的DCGN网络按照上述概念方案自动生成方法能自动生成相应的设计概念,结果如表2所示。自动生成的设计概念中提供了更加具体、可行的设计概念,例如:发明一种用于净化海水为饮用水的净化系统,或者使用太阳能来淡化海水,用于生产罐装饮用水或饮料制品。这些设计概念为沿海地区的居民或企业提供了产品开发早期的设计构想。1. The design problem of this case is to provide edible water sources for residents in coastal areas. In order to express the design problem more accurately and concisely, 10 graduate students majoring in mechanical engineering were invited to define the design problem using a limited number of keywords. Taking into account the advantages of abundant sunshine and light in coastal areas, the design team unanimously agreed to use "purification" (purification or purify), "desalination" (desalination or desalinate), "solar" (solar), "seawater" (seawater) and The "drink" keyword is used to define the design problem. By combining different keywords as design input, the corresponding design concept can be automatically generated by using the constructed DCGN network according to the above-mentioned concept scheme automatic generation method. The results are shown in Table 2. The automatically generated design concepts provide more specific and feasible design concepts, such as inventing a purification system for purifying seawater into drinking water, or using solar energy to desalinate seawater for the production of canned drinking water or beverage products. These design concepts provide residents or businesses in coastal areas with early design ideas for product development.
表2使用不同的设计问题关键词作为输入自动生成的概念方案Table 2 Automatically generated conceptual solutions using different design problem keywords as input
Figure PCTCN2022125347-appb-000082
Figure PCTCN2022125347-appb-000082
2、本发明所涉及的设计问题也可以由设计激励组成。在产品创新概念设计过程中,设计激励提供了丰富的、有价值的设计灵感,在传统的人工生成概念方案过程中,由设计激励到概念方案的往往是依赖设计师丰富的经验知识,并且生成概念方案的效率很低,对于缺乏经验的新手设计师该过程就变得非常困难。本实施例以获取的关于无人机的一些设计激励如表3所示。通过组合不同的 设计激励作为DCGN网络模型的输入,自动生成的概念方案如表4所示。由于组合种类很多,这里仅展示了一部分有价值的概念方案并进行分析。例如:2. The design problems involved in the present invention can also be composed of design incentives. In the process of product innovation concept design, design incentives provide rich and valuable design inspiration. In the traditional process of artificially generating conceptual solutions, the transition from design incentives to conceptual solutions often relies on the designer's rich experience and knowledge, and generates Conceptual solutions are inefficient and the process becomes very difficult for inexperienced novice designers. Some design incentives for UAVs obtained in this embodiment are shown in Table 3. By combining different design incentives as inputs to the DCGN network model, the automatically generated conceptual scheme is shown in Table 4. Since there are many types of combinations, only some valuable conceptual solutions are shown and analyzed here. For example:
(1)通过组合设计激励“drone”、“bio”、“radar”和“rescue”,DCGN网络自动生成了设计概念“a drone rescue radar system is disclosed that is capable of detecting the presence of an animal in the vicinity of the drone using bio”。(1) By combining the design incentives "drone", "bio", "radar" and "rescue", the DCGN network automatically generates the design concept "a drone rescue radar system is disclosed that is capable of detecting the presence of an animal in the "nearby of the drone using bio".
(2)通过组合设计激励“drone”、“fire”、“ground”和“data”,DCGN网络自动生成了设计概念“the drone may also be configured to receive ground fire data from the ground drone and to determine a location of the fire in response to detecting the resulting fire.”,以及“the drone may also be configured to receive ground fire data from the ground drone and to determine a location of the fire in response to determining the terrain”。(2) By combining the design incentives "drone", "fire", "ground" and "data", the DCGN network automatically generates the design concept "the drone may also be configured to receive ground fire data from the ground drone and to determine a location of the fire in response to detecting the resulting fire.", and "the drone may also be configured to receive ground fire data from the ground drone and to determine a location of the fire in response to determining the terrain".
表3无人机案例中检索的设计激励(不分先后顺序)Table 3 Design incentives retrieved in UAV cases (in no particular order)
Figure PCTCN2022125347-appb-000083
Figure PCTCN2022125347-appb-000083
表4无人机案例上组合不同设计激励自动生成的设计概念Table 4 Design concepts automatically generated by combining different design incentives on the UAV case
Figure PCTCN2022125347-appb-000084
Figure PCTCN2022125347-appb-000084
Figure PCTCN2022125347-appb-000085
Figure PCTCN2022125347-appb-000085
3、进一步地为了充分补充所涉及的设计问题,有一部分设计问题可以由设计需求来定义的。在产品设计的早期阶段,设计需求对于确定新产品的设计方向至关重要。在线产品评论数据为分析设计需求提供了准确、可靠和真实的信息,且易于访问。这里从某电商平台采用常规爬虫技术提取了某款奶瓶消毒柜的20918条用户评论文本,通过前面提供的数据预处理流程,分析得到了所包含的关键词和相应的词频,结果如表5所示。分析发现:用户主要表达了功能、消毒、容量、温度等方面的明确需求。为了应用设计需求得到设计问题,这里使用关键词“消毒”(disinfection或sterilization)、“温度”(temperature)、“功能”(function)、“容量”(capacity)作为DCGN网络模型的设计问题,自动生成的概念方案如表6所示。容易发现,使用不同的输入关键词组合,生成的概念方案也不相同。更重要的是,所有的自动生成的概念方案中均包含所输入的设计问题关键词,而且生成了一些可行的、有创意的概念方案,例如使用离子交换器(ion exchanger)来提升杀菌消毒能力,在一定程度上满足了设计需求。3. Further, in order to fully supplement the design issues involved, some design issues can be defined by design requirements. In the early stages of product design, design requirements are critical in determining the design direction of a new product. Online product review data provides accurate, reliable, authentic information for analyzing design needs, and is easily accessible. Here, we used conventional crawler technology to extract 20,918 user review texts of a certain milk bottle sterilizer from an e-commerce platform. Through the data preprocessing process provided above, we analyzed the included keywords and corresponding word frequencies. The results are shown in Table 5. shown. The analysis found that users mainly expressed clear needs in terms of functions, disinfection, capacity, temperature, etc. In order to apply the design requirements to obtain the design issues, the keywords "disinfection" (sterilization), "temperature" (temperature), "function" (function), and "capacity" (capacity) are used here as the design issues of the DCGN network model, automatically The generated conceptual scheme is shown in Table 6. It is easy to find that using different input keyword combinations will generate different conceptual solutions. More importantly, all automatically generated conceptual solutions contain the entered design problem keywords, and some feasible and creative conceptual solutions have been generated, such as using ion exchangers to improve sterilization and disinfection capabilities. , which meets the design needs to a certain extent.
表5在线用户评论数据中频率较高的前30个需求关键词Table 5 The top 30 most frequent demand keywords in online user review data
Figure PCTCN2022125347-appb-000086
Figure PCTCN2022125347-appb-000086
Figure PCTCN2022125347-appb-000087
Figure PCTCN2022125347-appb-000087
表6使用不同的设计需求关键词作为输入自动生成的设计概念Table 6 Design concepts automatically generated using different design requirement keywords as input
Figure PCTCN2022125347-appb-000088
Figure PCTCN2022125347-appb-000088
综上所述,如果由设计师对这些设计问题进行思考,仅仅依靠人工经验来产生概念方案,不仅难以创造出具有创新的概念方案,而且效率也会很低。本发明针对产品概念方案生成阶段以人工经验为主的跨领域设计知识迁移和重组难、自动生成设计概念方案难的问题,提出了基于设计概念生成网络(DCGN)的概念方案自动生成的方法,DCGN能够从海量的文本数据中自适应地学习跨领域设计知识的推理、迁移、重组等潜在的规律,并根据设计问题自动生成产品概念方案,不仅减轻了人工生成概念方案的负担,而且提高了设计效率,为智能化概念设计提供了新思路。To sum up, if designers think about these design issues and rely solely on artificial experience to generate conceptual solutions, it will not only be difficult to create innovative conceptual solutions, but the efficiency will also be very low. Aiming at the problem of difficulty in transferring and reorganizing cross-domain design knowledge and automatically generating design concept plans in the product concept plan generation stage, which is based on manual experience, the present invention proposes a method for automatically generating concept plans based on the Design Concept Generation Network (DCGN). DCGN can adaptively learn potential rules such as reasoning, migration, and reorganization of cross-domain design knowledge from massive text data, and automatically generate product concept solutions based on design problems. This not only reduces the burden of manually generating concept solutions, but also improves efficiency. Design efficiency provides new ideas for intelligent conceptual design.
本领域的普通技术人员将会意识到,这里所述的实施例是为了帮助读者理解本发明的原理,应被理解为本发明的保护范围并不局限于这样的特别陈述和实施例。本领域的普通技术人员可以根据本发明公开的这些技术启示做出各种不脱离本发明实质的其它各种具体变形和组合,这些变形和组合仍然在本发明的保护范围内。Those of ordinary skill in the art will appreciate that the embodiments described here are provided to help readers understand the principles of the present invention, and it should be understood that the scope of the present invention is not limited to such specific statements and embodiments. Those of ordinary skill in the art can make various other specific modifications and combinations based on the technical teachings disclosed in the present invention without departing from the essence of the present invention, and these modifications and combinations are still within the protection scope of the present invention.

Claims (10)

  1. 一种设计概念生成网络构建方法,其特征在于,该设计概念生成网络包括Transformer编码器,Transformer解码器,重要度约束矩阵生成模块,重要度约束嵌入层,跨注意力层和优化模块;所述设计概念生成网络构建方法包括以下步骤:A method for constructing a design concept generation network, characterized in that the design concept generation network includes a Transformer encoder, a Transformer decoder, an importance constraint matrix generation module, an importance constraint embedding layer, a cross-attention layer and an optimization module; The design concept generation network construction method includes the following steps:
    S1依据样本中输入词汇,利用Transformer编码器得到编码器隐藏层特征;S1 uses the Transformer encoder to obtain the hidden layer features of the encoder based on the input vocabulary in the sample;
    S2依据样本中目标序列,利用Transformer解码器得到解码器隐藏层特征;S2 uses the Transformer decoder to obtain the decoder hidden layer features based on the target sequence in the sample;
    S3依据样本中输入词汇和目标序列,利用重要度约束矩阵生成模块获取重要度约束矩阵;S3 uses the importance constraint matrix generation module to obtain the importance constraint matrix based on the input vocabulary and target sequence in the sample;
    S4利用重要度约束嵌入层将重要度约束矩阵映射至分布式向量空间,得到两个输入词汇重要度嵌入特征;S4 uses the importance constraint embedding layer to map the importance constraint matrix to the distributed vector space to obtain two input vocabulary importance embedding features;
    S5依据编码器隐藏层特征、解码器隐藏层特征和两个输入词汇重要度嵌入特征,利用跨注意力层获取生成序列;S5 uses the cross-attention layer to obtain the generated sequence based on the encoder hidden layer features, decoder hidden layer features and two input vocabulary importance embedding features;
    S6依据生成序列和目标序列构建损失函数,并基于损失函数利用优化模块对网络参数进行调整;然后重复步骤S1-S6,直至损失函数满足设定要求,得到设计概念生成网络。S6 constructs a loss function based on the generation sequence and the target sequence, and uses the optimization module to adjust the network parameters based on the loss function; then repeat steps S1-S6 until the loss function meets the set requirements, and the design concept generation network is obtained.
  2. 根据权利要求1所述的设计概念生成网络构建方法,其特征在于,步骤S1中,Transformer编码器按照以下公式得到编码器隐藏层特征h eThe design concept generation network construction method according to claim 1, characterized in that, in step S1, the Transformer encoder obtains the encoder hidden layer feature he according to the following formula:
    Figure PCTCN2022125347-appb-100001
    Figure PCTCN2022125347-appb-100001
    式中,x表示输入词汇;SA()表示空间注意力;
    Figure PCTCN2022125347-appb-100002
    分别表示Transformer编码器自注意力层的权重矩阵。
    In the formula, x represents the input vocabulary; SA() represents spatial attention;
    Figure PCTCN2022125347-appb-100002
    Represent the weight matrix of the Transformer encoder's self-attention layer respectively.
  3. 根据权利要求1所述的设计概念生成网络构建方法,其特征在于,步骤S2中,Transformer解码器通过自注意力层将前一时刻的目标序列y :t-1=[y 0,y 1,L,y t-1]映射至分布式的特征表示,即得到解码器隐藏层特征
    Figure PCTCN2022125347-appb-100003
    The design concept generation network construction method according to claim 1, characterized in that, in step S2, the Transformer decoder converts the target sequence y at the previous moment through the self-attention layer : t-1 = [y 0 , y 1 , L,y t-1 ] is mapped to the distributed feature representation, that is, the decoder hidden layer features are obtained
    Figure PCTCN2022125347-appb-100003
    Figure PCTCN2022125347-appb-100004
    Figure PCTCN2022125347-appb-100004
    式中,SA()表示空间注意力;
    Figure PCTCN2022125347-appb-100005
    分别表示Transformer编码器自注意力层的权重矩阵。
    In the formula, SA() represents spatial attention;
    Figure PCTCN2022125347-appb-100005
    Represent the weight matrix of the Transformer encoder's self-attention layer respectively.
  4. 根据权利要求3所述的设计概念生成网络构建方法,其特征在于,步骤S3中,用
    Figure PCTCN2022125347-appb-100006
    表示目标序列y :t中包含的输入词汇重要度约束向量,即C :t
    The design concept generation network construction method according to claim 3, characterized in that, in step S3, using
    Figure PCTCN2022125347-appb-100006
    Represents the input vocabulary importance constraint vector contained in the target sequence y :t , that is, C :t ;
    Figure PCTCN2022125347-appb-100007
    可计算为:
    Figure PCTCN2022125347-appb-100007
    Can be calculated as:
    Figure PCTCN2022125347-appb-100008
    Figure PCTCN2022125347-appb-100008
    式中,·表示向量或矩阵点积运算;
    Figure PCTCN2022125347-appb-100009
    为输入x在目标序列y :t中相对重要度向量,可计算为:
    In the formula, · represents the vector or matrix dot product operation;
    Figure PCTCN2022125347-appb-100009
    is the relative importance vector of input x in the target sequence y :t , which can be calculated as:
    Figure PCTCN2022125347-appb-100010
    Figure PCTCN2022125347-appb-100010
    式中,
    Figure PCTCN2022125347-appb-100011
    表示第i个输入词汇在目标序列y :t中相对重要度;w i表示第i个输入词汇在目标序列y :t中绝对重要度;w min表示输入词汇在目标序列y :t中的最小绝对重要度;w max表示输入词汇在目标序列y :t中的最大绝对重要度;[]为取整运算;M≥m>1,M表示整个训练用样本集中样本包含的最多输入词汇数目;
    In the formula,
    Figure PCTCN2022125347-appb-100011
    represents the relative importance of the i-th input vocabulary in the target sequence y :t ; w i represents the absolute importance of the i-th input vocabulary in the target sequence y: t ; w min represents the minimum value of the input vocabulary in the target sequence y :t Absolute importance; w max represents the maximum absolute importance of the input vocabulary in the target sequence y :t ; [] is the rounding operation; M≥m>1, M represents the maximum number of input vocabulary contained in the sample in the entire training sample set;
    Figure PCTCN2022125347-appb-100012
    表示目标序列y :t中包含的输入词汇约束;当目标序列y :t中包含了输入词汇的第i个词汇时,则向量c t中的第i个元素为1,可以通过如下方式计算:
    Figure PCTCN2022125347-appb-100012
    Indicates the input vocabulary constraints contained in the target sequence y :t ; when the target sequence y :t contains the i-th word of the input vocabulary, the i-th element in the vector c t is 1, which can be calculated as follows:
    Figure PCTCN2022125347-appb-100013
    Figure PCTCN2022125347-appb-100013
  5. 根据权利要求1或4所述的设计概念生成网络构建方法,其特征在于,步骤S4中,引入两个重要度约束嵌入矩阵
    Figure PCTCN2022125347-appb-100014
    Figure PCTCN2022125347-appb-100015
    将上述构建的重要度约束矩阵映射至分布式向量空间,得到两个输入词汇重要度嵌入特征
    Figure PCTCN2022125347-appb-100016
    Figure PCTCN2022125347-appb-100017
    在生成的第t个时刻,有:
    The design concept generation network construction method according to claim 1 or 4, characterized in that, in step S4, two importance constraint embedding matrices are introduced
    Figure PCTCN2022125347-appb-100014
    and
    Figure PCTCN2022125347-appb-100015
    Map the importance constraint matrix constructed above to the distributed vector space to obtain two input vocabulary importance embedding features.
    Figure PCTCN2022125347-appb-100016
    and
    Figure PCTCN2022125347-appb-100017
    At the t-th moment of generation, there is:
    Figure PCTCN2022125347-appb-100018
    Figure PCTCN2022125347-appb-100018
    Figure PCTCN2022125347-appb-100019
    Figure PCTCN2022125347-appb-100019
  6. 根据权利要求5所述的设计概念生成网络构建方法,其特征在于,步骤S5中,跨注意力层,融合处理编码器隐藏层特征h e、解码器隐藏层特征
    Figure PCTCN2022125347-appb-100020
    和两个输入词汇重要度嵌入特征
    Figure PCTCN2022125347-appb-100021
    得到当前时刻t的生成序列
    Figure PCTCN2022125347-appb-100022
    The design concept generation network construction method according to claim 5, characterized in that, in step S5, the encoder hidden layer features he and the decoder hidden layer features are fused across the attention layer
    Figure PCTCN2022125347-appb-100020
    and two input word importance embedding features
    Figure PCTCN2022125347-appb-100021
    Get the generated sequence at the current time t
    Figure PCTCN2022125347-appb-100022
    Figure PCTCN2022125347-appb-100023
    Figure PCTCN2022125347-appb-100023
    式中,
    Figure PCTCN2022125347-appb-100024
    表示解码器自注意力层的权重矩阵;
    In the formula,
    Figure PCTCN2022125347-appb-100024
    Represents the weight matrix of the decoder’s self-attention layer;
    在具体实现方式中,CA函数中第j个元素可表示为:In the specific implementation, the j-th element in the CA function can be expressed as:
    Figure PCTCN2022125347-appb-100025
    Figure PCTCN2022125347-appb-100025
    式中,
    Figure PCTCN2022125347-appb-100026
    Figure PCTCN2022125347-appb-100027
    i=0,1,…,M-1;j,l=0,1,…,d-1;() T表示转置运算。
    In the formula,
    Figure PCTCN2022125347-appb-100026
    Figure PCTCN2022125347-appb-100027
    i=0,1,…,M-1; j, l=0,1,…,d-1; () T represents the transpose operation.
  7. 根据权利要求1所述的设计概念生成网络构建方法,其特征在于,步骤S6中,对于给定的N个样本
    Figure PCTCN2022125347-appb-100028
    依据生成序列和目标序列构建的损失函数为:
    The design concept generation network construction method according to claim 1, characterized in that, in step S6, for the given N samples
    Figure PCTCN2022125347-appb-100028
    The loss function constructed based on the generated sequence and the target sequence is:
    Figure PCTCN2022125347-appb-100029
    Figure PCTCN2022125347-appb-100029
    式中,
    Figure PCTCN2022125347-appb-100030
    表示在t时刻的生成序列
    Figure PCTCN2022125347-appb-100031
    和对应时刻的目标序列y :t之间误差。
    In the formula,
    Figure PCTCN2022125347-appb-100030
    Represents the generated sequence at time t
    Figure PCTCN2022125347-appb-100031
    and the error between the target sequence y :t at the corresponding time.
  8. 一种概念方案自动生成方法,其特征在于,使用权利要求1至7任一方法构建的设计概念生成网络按照以下步骤执行操作:A method for automatically generating concept plans, characterized in that the design concept generation network constructed using any method of claims 1 to 7 performs operations in accordance with the following steps:
    L1,依据输入词汇,利用Transformer编码器得到编码器隐藏层特征;L1, based on the input vocabulary, use the Transformer encoder to obtain the hidden layer features of the encoder;
    L2,依据上一时刻生成序列,利用Transformer解码器得到当前时刻解码器隐藏层特征;L2, based on the sequence generated at the previous moment, use the Transformer decoder to obtain the hidden layer features of the decoder at the current moment;
    L3,依据样本中输入词汇和上一时刻生成序列,利用重要度约束矩阵生成模块获取重要度约束矩阵;L3, based on the input vocabulary in the sample and the generated sequence at the previous moment, use the importance constraint matrix generation module to obtain the importance constraint matrix;
    L4,利用重要度约束嵌入层将重要度约束矩阵映射至分布式向量空间,得到两个输入词汇重要度嵌入特征;L4, use the importance constraint embedding layer to map the importance constraint matrix to the distributed vector space, and obtain two input vocabulary importance embedding features;
    L5,依据编码器隐藏层特征、解码器隐藏层特征和两个输入词汇重要度嵌入特征,利用跨注意力层获取生成序列。L5, based on the encoder hidden layer features, decoder hidden layer features and two input vocabulary importance embedding features, use cross-attention layers to obtain the generated sequence.
  9. 根据权利要求8所述的概念方案自动生成方法,其特征在于,步骤L1中,输入词汇由来自于设计问题构建的关键词组成,或者由一个以上设计激励组成,或者由设计需求构建的关键词组成,或者至少两种输入词汇来源的组合。The method for automatically generating conceptual solutions according to claim 8, characterized in that, in step L1, the input vocabulary consists of keywords constructed from design problems, or consists of more than one design incentive, or keywords constructed from design requirements. composition, or a combination of at least two sources of input vocabulary.
  10. 根据权利要求8或9所述的概念方案自动生成方法,其特征在于,步骤L2中,依据上一时刻生成序列,按照以下公式计算得到当前时刻解码器隐藏层特征
    Figure PCTCN2022125347-appb-100032
    The method for automatically generating conceptual solutions according to claim 8 or 9, characterized in that, in step L2, the decoder hidden layer characteristics at the current moment are calculated according to the following formula based on the generated sequence at the previous moment.
    Figure PCTCN2022125347-appb-100032
    Figure PCTCN2022125347-appb-100033
    Figure PCTCN2022125347-appb-100033
    式中,y :t-1表示生成过程中t时刻解码器的输入序列,
    Figure PCTCN2022125347-appb-100034
    表示初始时刻给定的序列,
    Figure PCTCN2022125347-appb-100035
    表示前一时刻生成序列。
    In the formula, y :t-1 represents the input sequence of the decoder at time t during the generation process,
    Figure PCTCN2022125347-appb-100034
    Represents the sequence given at the initial moment,
    Figure PCTCN2022125347-appb-100035
    Represents the sequence generated at the previous moment.
PCT/CN2022/125347 2022-07-04 2022-10-14 Design concept generation network construction method and automatic concept scheme generation method WO2024007476A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US18/120,434 US20240005130A1 (en) 2022-07-04 2023-03-13 Method for constructing design concept generation network (dcgn) and method for automatically generating conceptual scheme

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210780085.4A CN115169227B (en) 2022-07-04 2022-07-04 Design concept generation network construction method and concept scheme automatic generation method
CN202210780085.4 2022-07-04

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US18/120,434 Continuation US20240005130A1 (en) 2022-07-04 2023-03-13 Method for constructing design concept generation network (dcgn) and method for automatically generating conceptual scheme

Publications (1)

Publication Number Publication Date
WO2024007476A1 true WO2024007476A1 (en) 2024-01-11

Family

ID=83490491

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/125347 WO2024007476A1 (en) 2022-07-04 2022-10-14 Design concept generation network construction method and automatic concept scheme generation method

Country Status (2)

Country Link
CN (1) CN115169227B (en)
WO (1) WO2024007476A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115169227B (en) * 2022-07-04 2023-07-07 四川大学 Design concept generation network construction method and concept scheme automatic generation method
CN115659852B (en) * 2022-12-26 2023-03-21 浙江大学 Layout generation method and device based on discrete potential representation

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107622047A (en) * 2017-09-04 2018-01-23 北京航空航天大学 A kind of extraction of design decision knowledge and expression
US20210334475A1 (en) * 2020-04-24 2021-10-28 Microsoft Technology Licensing, Llc Efficient transformer language models with disentangled attention and multi-step decoding
US20210365635A1 (en) * 2020-05-22 2021-11-25 Capital One Services, Llc Joint intent and entity recognition using transformer models
CN115169227A (en) * 2022-07-04 2022-10-11 四川大学 Design concept generation network construction method and concept scheme automatic generation method

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109062937B (en) * 2018-06-15 2019-11-26 北京百度网讯科技有限公司 The method of training description text generation model, the method and device for generating description text
CN110442705B (en) * 2019-04-23 2021-10-12 北京理工大学 Abstract automatic generation method based on concept pointer network
CN110390103B (en) * 2019-07-23 2022-12-27 中国民航大学 Automatic short text summarization method and system based on double encoders
CN110532328B (en) * 2019-08-26 2023-04-07 哈尔滨工程大学 Text concept graph construction method
CN110619043A (en) * 2019-08-30 2019-12-27 西南电子技术研究所(中国电子科技集团公司第十研究所) Automatic text abstract generation method based on dynamic word vector
CN112347242A (en) * 2020-11-11 2021-02-09 北京沃东天骏信息技术有限公司 Abstract generation method, device, equipment and medium
CN112434525A (en) * 2020-11-24 2021-03-02 平安科技(深圳)有限公司 Model reasoning acceleration method and device, computer equipment and storage medium
CN113128206B (en) * 2021-04-26 2022-04-19 中国科学技术大学 Question generation method based on word importance weighting
CN114089834A (en) * 2021-12-27 2022-02-25 杭州电子科技大学 Electroencephalogram identification method based on time-channel cascade Transformer network

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107622047A (en) * 2017-09-04 2018-01-23 北京航空航天大学 A kind of extraction of design decision knowledge and expression
US20210334475A1 (en) * 2020-04-24 2021-10-28 Microsoft Technology Licensing, Llc Efficient transformer language models with disentangled attention and multi-step decoding
US20210365635A1 (en) * 2020-05-22 2021-11-25 Capital One Services, Llc Joint intent and entity recognition using transformer models
CN115169227A (en) * 2022-07-04 2022-10-11 四川大学 Design concept generation network construction method and concept scheme automatic generation method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
ZHANG MIN; ZENG BIQING; HAN XULI; XU RUYANG;: "DAPC: Dual Attention and Pointer-Coverage Network Based Summarization Model", COMPUTER ENGINEERING AND APPLICATIONS, HUABEI JISUAN JISHU YANJIUSUO, CN, vol. 56, no. 08, 31 December 2020 (2020-12-31), CN , pages 149 - 157, XP009551750, ISSN: 1002-8331 *

Also Published As

Publication number Publication date
CN115169227B (en) 2023-07-07
CN115169227A (en) 2022-10-11

Similar Documents

Publication Publication Date Title
WO2024007476A1 (en) Design concept generation network construction method and automatic concept scheme generation method
CN107239444B (en) A kind of term vector training method and system merging part of speech and location information
CN104298651B (en) Biomedicine named entity recognition and protein interactive relationship extracting on-line method based on deep learning
CN105740349B (en) A kind of sensibility classification method of combination Doc2vec and convolutional neural networks
CN109766427B (en) Intelligent question-answering method based on collaborative attention for virtual learning environment
CN110334219A (en) The knowledge mapping for incorporating text semantic feature based on attention mechanism indicates learning method
CN106845411A (en) A kind of video presentation generation method based on deep learning and probability graph model
CN106682220A (en) Online traditional Chinese medicine text named entity identifying method based on deep learning
CN106650789A (en) Image description generation method based on depth LSTM network
CN110516245A (en) Fine granularity sentiment analysis method, apparatus, computer equipment and storage medium
CN106484681A (en) A kind of method generating candidate&#39;s translation, device and electronic equipment
CN110222163A (en) A kind of intelligent answer method and system merging CNN and two-way LSTM
CN103995805B (en) The word processing method of the big data of text-oriented
CN110795571A (en) Cultural tourism resource recommendation method based on deep learning and knowledge graph
CN112990296A (en) Image-text matching model compression and acceleration method and system based on orthogonal similarity distillation
Zhao et al. Multi-task learning with graph attention networks for multi-domain task-oriented dialogue systems
He et al. A modularized architecture of multi-branch convolutional neural network for image captioning
CN116522945A (en) Model and method for identifying named entities in food safety field
Li et al. Multimodal fusion with co-attention mechanism
Ao et al. Tpenas: A two-phase evolutionary neural architecture search for remote sensing image classification
CN103413125B (en) Terrified video frequency identifying method based on identification example selection multi-instance learning
CN107622047A (en) A kind of extraction of design decision knowledge and expression
CN107506351A (en) It is a kind of that special Semantic Similarity analysis method is pushed away based on character convolutional network
CN117093692A (en) Multi-granularity image-text matching method and system based on depth fusion
CN113568901A (en) Data processing method, device and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22950036

Country of ref document: EP

Kind code of ref document: A1