WO2023134084A1 - Multi-label identification method and apparatus, electronic device, and storage medium - Google Patents

Multi-label identification method and apparatus, electronic device, and storage medium Download PDF

Info

Publication number
WO2023134084A1
WO2023134084A1 PCT/CN2022/090726 CN2022090726W WO2023134084A1 WO 2023134084 A1 WO2023134084 A1 WO 2023134084A1 CN 2022090726 W CN2022090726 W CN 2022090726W WO 2023134084 A1 WO2023134084 A1 WO 2023134084A1
Authority
WO
WIPO (PCT)
Prior art keywords
vector
label
comment
user
matrix
Prior art date
Application number
PCT/CN2022/090726
Other languages
French (fr)
Chinese (zh)
Inventor
舒畅
陈又新
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2023134084A1 publication Critical patent/WO2023134084A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Definitions

  • the present application relates to the technical field of artificial intelligence, and in particular to a multi-tag identification method, device, electronic equipment and storage medium.
  • the main purpose of the embodiment of the present application is to provide a multi-tag identification method, device, electronic device and storage medium, aiming at improving the recognition accuracy and recognition efficiency of user portrait tags.
  • the embodiment of the present application proposes a multi-label identification method, the method comprising:
  • the raw data includes user basic data, user behavior data and user comment data
  • the comment text word segment vector is input into the pre-trained comparative learning model, so that the comment text word segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied to obtain the comment word embedding vector;
  • the target portrait label is obtained.
  • the embodiment of the present application proposes a multi-tag identification device, which includes:
  • a data acquisition module configured to acquire raw data, wherein the raw data includes user basic data, user behavior data and user comment data;
  • a normalization module configured to perform normalization processing on the user basic data to obtain user basic features
  • Feature extraction module for carrying out feature extraction to described user behavior data by pre-trained graph convolution model, obtains behavior feature matrix
  • a word segmentation module configured to perform word segmentation processing on the user comment data to obtain a comment text word segment vector
  • a comparative learning module for inputting the comment text word segment vector into a pre-trained comparative learning model, so that the comment text word segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied, Get the comment embedding vector;
  • the fusion module is used to perform fusion processing on the user basic features, the behavior feature matrix and the comment word embedding vector to obtain a standard portrait feature vector;
  • a tag recognition module configured to perform tag recognition processing on the standard portrait feature vector through a pre-trained tag recognition model to obtain the probability value of each preset portrait tag;
  • the comparison module is used to obtain the target portrait label according to the magnitude relationship between the probability value and the preset probability threshold.
  • the embodiment of the present application provides an electronic device, the electronic device includes a memory, a processor, a program stored in the memory and operable on the processor, and a program for implementing the processor
  • a data bus connecting and communicating with the memory when the program is executed by the processor, a multi-label identification method is implemented, wherein the multi-label identification method includes: obtaining original data, wherein the original The data includes user basic data, user behavior data, and user comment data; the user basic data is normalized to obtain user basic features; the user behavior data is extracted through a pre-trained graph convolution model to obtain Behavior feature matrix; word segmentation processing is carried out to described user comment data, obtain comment text phrase vector; Described comment text phrase vector is input in the comparative learning model of pre-training, so that described comment text phrase vector and all The reference word embedding matrix in the comparative learning model is multiplied by matrix to obtain the comment word embedding vector; the user basic feature, the behavior feature matrix and the comment word embedding vector are fused to obtain a standard
  • the embodiment of the present application provides a storage medium, the storage medium is a computer-readable storage medium for computer-readable storage, the storage medium stores one or more programs, and the one or more This program can be executed by one or more processors to implement a multi-label identification method, wherein the multi-label identification method includes: obtaining original data, wherein the original data includes user basic data, user behavior data, and User comment data; normalize the user basic data to obtain user basic features; perform feature extraction on the user behavior data through a pre-trained graph convolution model to obtain a behavior feature matrix; Carry out participle processing, obtain comment text phrase vector; Described comment text phrase vector is input in the comparative learning model of pre-training, so that described comment text phrase vector and the reference word embedding matrix in described contrast learning model Perform matrix multiplication to obtain comment word embedding vectors; perform fusion processing on the user basic features, the behavior feature matrix, and the comment word embedding vectors to obtain standard portrait feature vectors; The standard portrait feature vector performs label recognition processing
  • the multi-tag recognition method, device, electronic equipment, and storage medium proposed in this application can greatly shorten the model training time, improve the recognition efficiency, and improve the recognition accuracy of user portrait tags.
  • Fig. 1 is a flow chart of the multi-tag identification method provided by the embodiment of the present application.
  • Fig. 2 is the flowchart of step S103 in Fig. 1;
  • Fig. 3 is the flowchart of step S104 in Fig. 1;
  • Fig. 4 is the flowchart of step S105 in Fig. 1;
  • Fig. 5 is another flow chart of the multi-tag identification method provided by the embodiment of the present application.
  • Fig. 6 is the flowchart of step S107 in Fig. 1;
  • Fig. 7 is the flowchart of step S108 in Fig. 1;
  • Fig. 8 is a schematic structural diagram of a multi-tag identification device provided by an embodiment of the present application.
  • FIG. 9 is a schematic diagram of a hardware structure of an electronic device provided by an embodiment of the present application.
  • embodiments of the present application provide a multi-tag identification method, device, electronic equipment, and storage medium, aiming at improving the identification accuracy of user portrait tags.
  • the multi-tag identification method, device, electronic device, and storage medium provided in the embodiments of the present application are specifically described through the following embodiments. First, the multi-tag identification method in the embodiment of the present application is described.
  • AI artificial intelligence
  • the embodiments of the present application may acquire and process relevant data based on artificial intelligence technology.
  • artificial intelligence is the theory, method, technology and application system that uses digital computers or machines controlled by digital computers to simulate, extend and expand human intelligence, perceive the environment, acquire knowledge and use knowledge to obtain the best results. .
  • Artificial intelligence basic technologies generally include technologies such as sensors, dedicated artificial intelligence chips, cloud computing, distributed storage, big data processing technology, operation/interaction systems, and mechatronics.
  • Artificial intelligence software technology mainly includes computer vision technology, robotics technology, biometrics technology, speech processing technology, natural language processing technology, and machine learning/deep learning.
  • the multi-tag identification method provided in the embodiment of the present application relates to the fields of artificial intelligence and digital medical technology.
  • the multi-tag identification method provided in the embodiment of the present application can be applied to a terminal, can also be applied to a server, and can also be software running on a terminal or a server.
  • the terminal can be a smart phone, a tablet computer, a notebook computer, a desktop computer, etc.
  • the server end can be configured as an independent physical server, or can be configured as a server cluster or a distributed system composed of multiple physical servers, or It can be configured as a cloud that provides basic cloud computing services such as cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communications, middleware services, domain name services, security services, CDN, and big data and artificial intelligence platforms.
  • the server; the software may be an application to realize the multi-label identification method, but is not limited to the above forms.
  • the application can be used in numerous general purpose or special purpose computer system environments or configurations. Examples: personal computers, server computers, handheld or portable devices, tablet-type devices, multiprocessor systems, microprocessor-based systems, set-top boxes, programmable consumer electronics, network PCs, minicomputers, mainframe computers, including A distributed computing environment for any of the above systems or devices, etc.
  • This application may be described in the general context of computer-executable instructions, such as program modules, being executed by a computer.
  • program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types.
  • the application may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network.
  • program modules may be located in both local and remote computer storage media including storage devices.
  • Fig. 1 is an optional flow chart of the multi-tag identification method provided by the embodiment of the present application.
  • the method in Fig. 1 may include but not limited to steps S101 to S108.
  • Step S101 obtaining original data, wherein the original data includes user basic data, user behavior data and user comment data;
  • Step S102 performing normalization processing on the basic user data to obtain basic user features
  • Step S103 using the pre-trained graph convolution model to perform feature extraction on user behavior data to obtain a behavior feature matrix
  • Step S104 performing word segmentation processing on user comment data to obtain comment text word segment vectors
  • Step S105 inputting the comment text word segment vector into the pre-trained comparative learning model, so that the comment text word segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied to obtain the comment word embedding vector;
  • Step S106 performing fusion processing on user basic features, behavior feature matrix and comment word embedding vectors to obtain standard portrait feature vectors
  • Step S107 using the pre-trained label recognition model to perform label recognition processing on the standard portrait feature vector to obtain the probability value of each preset portrait label;
  • Step S108 according to the magnitude relationship between the probability value and the preset probability threshold, the target portrait label is obtained.
  • the multi-label recognition method of the present application can realize the recognition of different portrait labels through a label recognition model. Compared with the traditional technology, which needs to train classifiers for different label categories separately, the model can be greatly shortened. The training time improves the recognition efficiency. At the same time, the multi-label recognition method of the present application performs corresponding data preprocessing for different types of user data, so that the obtained standard portrait feature vectors can better meet the recognition requirements, and can improve the recognition accuracy of user portrait labels.
  • the original user data can be crawled from multiple preset data sources by means of a web crawler, wherein the basic data includes the user's gender, education background, age group, etc.; behavior data Including the user's click data on the display of the course content and the click data of the recommended courses on the course page, etc.; the comment data is the user's text comment data on the course, etc.
  • a series of numbers can be set for different types of basic data according to preset normalization conditions.
  • Educational background is divided into ⁇ 1,2,3,4,5,6,7,8 ⁇ set, 1 represents primary school, 2 represents junior high school, 3 represents technical secondary school, 4 represents high school, 5 represents junior college, 6 represents undergraduate, 7 represents Master, 8 represents Ph.D.
  • the age group is divided into ⁇ 5, 6, 7, 8, 9, 0 ⁇ sets, 5 represents the post-50s, 6 represents the post-60s, and so on.
  • step S103 may include but not limited to include steps S201 to S204:
  • Step S201 mapping user behavior data to a preset vector space to obtain user behavior feature vectors
  • Step S202 constructing a behavior feature map according to the preset course types and user behavior feature vectors
  • Step S203 performing graph convolution processing on the behavior feature map to obtain a behavior degree matrix and a behavior adjacency matrix
  • Step S204 performing difference processing on the behavior degree matrix and the behavior matrix to obtain a behavior characteristic matrix.
  • the user behavior data can be mapped to a preset vector space by using the MLP network to map the semantic space of the user behavior data to the vector space, and obtain the user behavior feature vector.
  • step S202 record each preset course as a node, analyze the user's behavior data, and if it is detected that the user clicks on another course through the recommendation module of a course page, then establish a link between the two courses. side.
  • the relationship between each course type and user behavior feature vector is constructed to obtain an undirected graph, which is the behavior feature graph.
  • the adjacency matrix is 0; therefore, by performing graph convolution processing on the behavior feature map, the Laplace transform of the behavior feature map can be realized, and the behavior degree matrix (ie, the angle matrix D) and the behavior adjacency matrix (ie, the adjacency matrix A) can be obtained.
  • y is the output value and ⁇ is the sigmoid activation function.
  • L is the Laplacian matrix
  • x is the input labeled behavioral feature map
  • j is the number of rows of the Laplacian matrix, where j is generally much smaller than the number of nodes in the behavioral feature map.
  • is the weight matrix.
  • the parameter values of the weight matrix are randomly generated when the graph convolution model is initialized, and can be adjusted later by training the graph convolution model. Specifically, calculate the error between the marked behavior features and the predicted features, and then The error is backpropagated to update the parameter value to optimize the graph convolution model.
  • step S104 may include but not limited to include steps S301 to S302:
  • Step S301 performing word segmentation processing on the user comment data through a preset word segmenter to obtain comment text segments
  • Step S302 encoding the word segment of the review text to obtain a word segment vector of the review text.
  • step S301 when using the Jieba tokenizer to perform word segmentation processing on the user comment data, first generate a directed acyclic graph corresponding to the user comment data by comparing the dictionary in the Jieba tokenizer, and then according to the preset selection Patterns and dictionaries search for the shortest path on the directed acyclic graph, and intercept the user comment data according to the shortest path, or directly intercept the user comment data to obtain comment text segments.
  • HMM Hidden Markov Model
  • the positions B, M, E, and S of the characters in the comment text segment are regarded as hidden states, and the characters are observed states, where B/M/E/S represent the words that appear at the beginning, middle, end, and word respectively into words.
  • a dictionary file is used to store representation probability matrix, initial probability vector and transition probability matrix between characters respectively. Then use the Viterbi algorithm to solve the maximum possible hidden state, so as to obtain the comment text segment.
  • a preset BERT encoder may be used to encode the word segment of the comment text, so that each character on the word segment of the comment text carries a corresponding code, thereby obtaining a word segment vector of the comment text.
  • the method before step S105, the method further includes pre-training a contrastive learning model, which may specifically include but not limited to steps a to f:
  • sample pairs according to the initial embedding data, wherein the sample pairs include positive example pairs and negative example pairs;
  • step a and step b first obtain the sample user data, encode the sample user data, map the sample user data to the embedding space, and perform vector representation on the sample user data, so as to obtain the initial embedded data (ie Initial embedding data), the initial embedding data includes positive sample data and negative sample data.
  • Initial embedding data the initial embedding data includes positive sample data and negative sample data.
  • the data enhancement processing is performed on the initial embedded data through the dropout mask mechanism; the embodiment of the present application replaces the traditional data enhancement method through the dropout mask mechanism, that is, the same sample data is input into the dropout encoder twice to obtain
  • the two vectors of the two vectors are used as positive example pairs for comparative learning, and the effect is good enough, because for example, a different dropout mask is randomly generated for each dropout inside BERT, so only the same sample data (that is, the initial Embedding data) is input to the simCSE model twice, and the two vectors obtained are the results of applying two different dropout masks.
  • the dropout mask is a random network model, which is the mask of the model parameter W, which prevents overfitting.
  • the data (that is, the first vector and the second vector) obtained through data enhancement processing are positive example pairs, and other data that have not undergone data enhancement are negative example pairs.
  • some of the initial embedded data in a batch can be processed through data enhancement to obtain positive example pairs, and the other part of the initial embedded data can be used as negative example pairs.
  • step d is performed to input the sample pair into the contrastive learning model.
  • both the first similarity and the second similarity are cosine similarity.
  • step f may include, but is not limited to include:
  • the loss function Maximize the first similarity to the first value and minimize the second similarity to the first value to optimize the loss function; where the first similarity is the numerator of the loss function, the first similarity and the second The similarity is the denominator of the loss function, the first value is 1, and the second value is 0.
  • the numerator is the first similarity of the corresponding positive example pair
  • the denominator is the first similarity and the second similarity of all negative example pairs
  • the value of the molecular formula composed of the numerator and the denominator is wrapped in -log() In this way, the loss function can be minimized by maximizing the numerator and minimizing the denominator.
  • minimizing the loss function infoNCE loss is to maximize the numerator and minimize the denominator, that is, to maximize the first similarity of the positive pair and minimize the second similarity of the negative pair, and the loss The function is minimized to realize the optimization of the loss function. More specifically, the loss function is shown in formula (2):
  • the loss function represents the loss of sample i; in the loss function, the numerator is the similarity of the positive pair , the denominator is the similarity between positive pairs and all negative pairs, and then wrap this value in -log(), so that maximizing the numerator and minimizing the denominator can minimize the loss function.
  • step S105 may include but not limited to include steps S401 to S402:
  • Step S401 input the comment text word segment vector into the comparative learning model, so that the comment text word segment vector and the reference word embedding matrix are matrix multiplied to obtain a plurality of basic word embedding vectors;
  • Step S402 performing mapping processing on the basic word embedding vectors to obtain review word embedding vectors.
  • step S401 is executed, and the value of the reference word embedding matrix in the comparison model can be completely fixed by training the comparison model, and other model parameters of the comparison model are also fixed. Therefore, inputting the comment text word segment vector into the comparison model, the fixed reference word embedding matrix can be used to perform matrix multiplication with each comment text word segment vector to obtain multiple basic word embedding vectors.
  • step S402 use the fixed MLP network in the comparison model to perform mapping processing on the basic word embedding vector to obtain the comment word embedding vector.
  • the MLP network includes a linear layer, a ReLu activation function, and a linear layer.
  • step S106 when step S106 is executed, the basic feature data and the behavioral feature matrix are respectively vectorized to obtain the basic feature vector and the behavioral feature vector, and then the basic feature vector, the behavioral feature vector and the word embedding feature vector are performed Fusion processing to get the standard feature vector.
  • the standard feature vector X [gender, education, age group, [GCN], [BERT]].
  • GCN is a 256-dimensional vector
  • BERT is a 512-dimensional vector
  • X is a 3+256+512 vector.
  • the method further includes pre-training the label recognition model, which may specifically include but not limited to steps S501 to S505:
  • Step S501 acquiring marked user data
  • Step S502 performing feature extraction on the labeled user data to obtain sample feature vectors
  • Step S503 inputting the sample feature vector into the label recognition model
  • Step S504 calculate the sample probability prediction value of each portrait label category through the loss function of the label recognition model
  • Step S505 optimize the loss function of the label recognition model according to the sample probability prediction value, so as to update the label recognition model.
  • the label recognition model can be a textcnn model, and the label recognition model includes an Embedding layer, a convolution layer, a pooling layer and an output layer.
  • the Embedding layer of the label recognition model can use ELMO, GLOVE, Word2Vector, Bert and other algorithms to generate a dense vector from the input text data. Then, the dense vector is convoluted and pooled through the convolution layer and pooling layer of the label recognition model to obtain the target feature vector, and then the target feature vector is input to the output layer, and the preset function in the output layer is The classification operation can be performed on the target feature vector to obtain the label feature vector and the probability value of each preset category.
  • step S501 is executed to obtain marked user data, where the marked user data includes user portrait category labels. Furthermore, step S502 is executed, using the MLP network to perform multiple mapping processes on the labeled user data to obtain sample feature vectors.
  • step S503 is executed to input the sample feature vector into the label recognition model.
  • step S504 the sample feature vector is generated into a dense feature vector through the Embedding layer of the label recognition model, and then the dense feature vector is convoluted and pooled through the convolution layer and the pooling layer to obtain the target The feature vector, and then input the target feature vector to the output layer, and calculate the sample probability prediction value of each portrait label category through the loss function; where the loss function is shown in formula (3):
  • t is the target value (target), and t needs to take a value between [0,1]. Since in the embodiment of this application, t is used as a portrait label category, the value of t is 0 or 1, and o indicates label identification The model's probabilistic predictions.
  • step S505 is executed to calculate the model loss of the label recognition model based on the sample probability prediction value, that is, the loss value, and then use the gradient descent method to backpropagate the loss value, feed the loss value back to the label recognition model, and modify the label recognition model.
  • the preset iteration condition is that the number of iterations can reach the preset value, or the variance of the loss function is smaller than the preset threshold.
  • the backpropagation can be stopped, and the final model parameter can be used as the final model parameter to complete the update of the label recognition model.
  • step S107 may also include but not limited to steps S601 to S602:
  • Step S601 reconstructing the standard portrait feature vector according to the preset label dimension to obtain the label feature vector
  • Step S602 using a preset function to identify the tag feature vector to obtain the probability value of each preset portrait tag.
  • step S601 is first performed to reconstruct the standard image feature vector according to the preset label dimension and encoder, for example, encode the standard image feature vector according to the bottom-up encoding sequence and label dimension.
  • the standard portrait feature vector is encoded for the first time to obtain the bottom label feature vector z1, and then the downsampling process is performed layer by layer to obtain the label feature vector [z2, z3..., zk] corresponding to each label dimension.
  • step S602 the preset function is a sigmoid function, and the sigmoid function can be expressed as shown in formula (4):
  • the sigmoid function will classify the label feature vector according to the preset portrait label category, and create a probability distribution on each portrait label category, so as to obtain the value of each preset portrait label. probability value.
  • step S108 may also include but not limited to include steps S701 to S702:
  • Step S701 including the portrait labels whose probability value is greater than or equal to the preset probability threshold into the same set to obtain a set of candidate portrait labels;
  • Step S702 screening the candidate portrait label set to obtain the target portrait label.
  • first execute step S701 if the probability value is less than the preset probability threshold, filter out the portrait label corresponding to the probability value; if the probability value is greater than or equal to the preset probability threshold, then include the portrait label corresponding to the probability value into the candidate Portrait labels set.
  • the preset probability threshold is 0.6, and when the probability value is greater than or equal to 0.6, it can be considered that the user has the current portrait tag.
  • the portrait tags in the candidate portrait tag set can be screened by means of manual review, etc., and the portrait tag with the highest matching degree with the current user is extracted to obtain the target portrait tag.
  • the portrait tags in the candidate portrait tag set can also be arranged in descending order according to the size of the probability value, and the top five portrait tags are selected as the target portrait tags of the current user.
  • Other methods may also be used to filter the portrait tags in the candidate portrait tag set, which is not limited thereto.
  • raw data is acquired, wherein the raw data includes basic user data, user behavior data, and user comment data.
  • the user basic data is normalized to obtain the user basic features;
  • the user behavior data is extracted through the pre-trained graph convolution model to obtain the behavior feature matrix;
  • the user comment data is word-segmented to obtain the comment text word segment vector, and input the comment text segment vector into the pre-trained comparative learning model, so that the review text segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied to obtain the review word embedding vector.
  • different types of data can be preprocessed separately to obtain user basic features, behavior feature matrices, and comment embedding vectors, which improves the rationality of user data.
  • the standard portrait feature vector is obtained by fusing the user's basic features, behavior feature matrix, and comment word embedding vector.
  • the pre-trained label recognition model is used to perform label recognition processing on the standard portrait feature vector to obtain the probability value of each preset portrait label, and obtain the target portrait label according to the relationship between the probability value and the preset probability threshold.
  • the multi-label recognition method of the present application can realize the recognition of different portrait labels through a label recognition model. Compared with the traditional technology that needs to train classifiers for different label categories separately, it can greatly shorten the model training time and improve the recognition efficiency.
  • the multi-label recognition method of the present application performs corresponding data preprocessing for different types of user data, so that the obtained standard portrait feature vectors can better meet the recognition requirements, and can improve the recognition accuracy of user portrait labels.
  • the embodiment of the present application also provides a multi-label identification device, which can realize the above-mentioned multi-label identification method, the device includes:
  • a data acquisition module 801, configured to acquire raw data, wherein the raw data includes user basic data, user behavior data and user comment data;
  • a normalization module 802 configured to perform normalization processing on user basic data to obtain user basic features
  • the feature extraction module 803 is used to perform feature extraction on user behavior data through a pre-trained graph convolution model to obtain a behavior feature matrix
  • the word segmentation module 804 is used to perform word segmentation processing on the user comment data to obtain the comment text word segment vector;
  • Contrastive learning module 805 for inputting the comment text word segment vector into the pre-trained comparative learning model, so that the comment text word segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied to obtain the comment word embedding vector ;
  • the fusion module 806 is used to fuse the user basic features, behavior feature matrix and comment word embedding vectors to obtain standard portrait feature vectors;
  • the label recognition module 807 is used to perform label recognition processing on the standard portrait feature vector through the pre-trained label recognition model to obtain the probability value of each preset portrait label;
  • the comparison module 808 is configured to obtain the target portrait label according to the magnitude relationship between the probability value and the preset probability threshold.
  • the specific implementation of the multi-tag identification device is basically the same as the specific embodiment of the above-mentioned multi-label identification method, and will not be repeated here.
  • the embodiment of the present application also provides an electronic device, the electronic device includes: a memory, a processor, a program stored in the memory and operable on the processor, and a data bus for realizing connection and communication between the processor and the memory , when the program is executed by the processor, the above multi-label identification method is realized.
  • the electronic device may be any intelligent terminal including a tablet computer, a vehicle-mounted computer, and the like.
  • FIG. 9 illustrates a hardware structure of an electronic device in another embodiment.
  • the electronic device includes:
  • the processor 901 may be implemented by a general-purpose CPU (Central Processing Unit, central processing unit), a microprocessor, an application-specific integrated circuit (Application Specific Integrated Circuit, ASIC), or one or more integrated circuits, and is used to execute related programs, so as to realize The technical solutions provided by the embodiments of the present application;
  • a general-purpose CPU Central Processing Unit, central processing unit
  • a microprocessor an application-specific integrated circuit (Application Specific Integrated Circuit, ASIC), or one or more integrated circuits, and is used to execute related programs, so as to realize The technical solutions provided by the embodiments of the present application;
  • ASIC Application Specific Integrated Circuit
  • the memory 902 may be implemented in the form of a read-only memory (ReadOnlyMemory, ROM), a static storage device, a dynamic storage device, or a random access memory (RandomAccessMemory, RAM).
  • the memory 902 can store operating systems and other application programs.
  • the relevant program codes are stored in the memory 902 and called by the processor 901 to execute a multi- The label recognition method, wherein the multi-label recognition method includes: obtaining original data, wherein the original data includes user basic data, user behavior data, and user comment data; normalizing the user basic data to obtain user basic features;
  • the trained graph convolution model extracts features from user behavior data to obtain a behavior feature matrix; performs word segmentation processing on user comment data to obtain comment text segment vectors; input comment text segment vectors into the pre-trained comparative learning model, Multiply the word segment vector of the comment text with the reference word embedding matrix in the comparative learning model to obtain the comment word embedding vector; perform fusion processing on the user basic features, behavior feature matrix and comment word embedding vector to obtain the standard portrait feature vector Carrying out label recognition processing on the standard portrait feature vector through the pre-trained label recognition model to obtain the probability value of each preset portrait label; according to the size relationship between the probability value
  • the input/output interface 903 is used to realize information input and output
  • the communication interface 904 is used to realize the communication interaction between the device and other devices, and the communication can be realized through a wired method (such as USB, network cable, etc.), or can be realized through a wireless method (such as a mobile network, WIFI, Bluetooth, etc.);
  • bus 905 for transferring information between various components of the device (such as processor 901, memory 902, input/output interface 903 and communication interface 904);
  • the processor 901 , the memory 902 , the input/output interface 903 and the communication interface 904 are connected to each other within the device through the bus 905 .
  • An embodiment of the present application also provides a storage medium, which is a computer-readable storage medium for computer-readable storage, and the computer-readable storage medium may be non-volatile or volatile.
  • the storage medium stores one or more programs, and the one or more programs can be executed by one or more processors to implement a multi-label identification method, wherein the multi-label identification method includes: obtaining original data, wherein the original data Including user basic data, user behavior data and user comment data; normalize user basic data to obtain user basic features; use pre-trained graph convolution model to extract user behavior data to obtain behavior feature matrix;
  • the user comment data is subjected to word segmentation processing to obtain the comment text segment vector; the comment text segment vector is input into the pre-trained comparative learning model, so that the comment text segment vector and the reference word embedding matrix in the comparative learning model are matrix correlated.
  • memory can be used to store non-transitory software programs and non-transitory computer-executable programs.
  • the memory may include high-speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid-state storage devices.
  • the memory optionally includes memory located remotely from the processor, and these remote memories may be connected to the processor via a network. Examples of the aforementioned networks include, but are not limited to, the Internet, intranets, local area networks, mobile communication networks, and combinations thereof.
  • the device embodiments described above are only illustrative, and the units described as separate components may or may not be physically separated, that is, they may be located in one place, or may be distributed to multiple network units. Part or all of the modules can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit.
  • the above-mentioned integrated units can be implemented in the form of hardware or in the form of software functional units.
  • the integrated unit is realized in the form of a software function unit and sold or used as an independent product, it can be stored in a computer-readable storage medium.
  • the technical solution of the present application is essentially or part of the contribution to the prior art or all or part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium , including multiple instructions to make a computer device (which may be a personal computer, a server, or a network device, etc.) execute all or part of the steps of the method in each embodiment of the present application.
  • the aforementioned storage media include: U disk, mobile hard disk, read-only memory (Read-Only Memory, referred to as ROM), random access memory (Random Access Memory, referred to as RAM), magnetic disk or optical disc, etc., which can store programs. medium.
  • ROM read-only memory
  • RAM random access memory
  • magnetic disk or optical disc etc., which can store programs. medium.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • General Engineering & Computer Science (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Image Analysis (AREA)

Abstract

The present application relates to the technical field of artificial intelligence. Embodiments of the present application provide a multi-label identification method and apparatus, an electronic device, and a storage medium. The method comprises: performing normalization processing on user basic data to obtain user basic features; performing feature extraction on user behavior data by means of a graph convolution model to obtain a behavior feature matrix; performing word segmentation processing on user comment data to obtain comment text word segment vectors; inputting the comment text word segment vectors to a comparative learning model so that matrix multiplication is performed on the comment text word segment vectors and a reference word embedding matrix so as to obtain comment word embedding vectors; performing fusion processing on the user basic features, the behavior feature matrix, and the comment word embedding vectors to obtain standard portrait feature vectors; performing label identification processing on the standard portrait feature vectors by means of a label identification model to obtain probability values of portrait labels; and obtaining a target portrait label according to the probability values. According to embodiments of the present application, the accuracy with which portrait labels of a user are identified is improved.

Description

多标签识别方法、装置、电子设备及存储介质Multi-label identification method, device, electronic equipment and storage medium
本申请要求于2022年1月11日提交中国专利局、申请号为202210027793.0,发明名称为“多标签识别方法、装置、电子设备及存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application with the application number 202210027793.0 filed on January 11, 2022, and the title of the invention is "multi-label identification method, device, electronic equipment and storage medium", the entire content of which is incorporated by reference incorporated in this application.
技术领域technical field
本申请涉及人工智能技术领域,尤其涉及一种多标签识别方法、装置、电子设备及存储介质。The present application relates to the technical field of artificial intelligence, and in particular to a multi-tag identification method, device, electronic equipment and storage medium.
背景技术Background technique
目前,在对互联网用户进行画像标签时,常常采用人工标注或者机器学习的方式来对画像标签进行识别和分类。现有方法采用人工标注方式时,往往需要经过长时间的标记处理,且出错率较高,影响识别准确性;而当采用机器学习的方式对多标签画像进行识别时,往往需要针对不同的标签类别,分别训练分类器,往往需要花费较多的时间进行模型训练,影响识别效率。因此,如何提供一种多标签识别方法,能够提高用户画像标签的识别准确性及识别效率,成为了亟待解决的技术问题。At present, when labeling portraits of Internet users, manual labeling or machine learning is often used to identify and classify the portrait labels. When the existing method uses manual labeling, it often needs to go through a long period of labeling processing, and the error rate is high, which affects the accuracy of recognition; and when using machine learning to identify multi-label portraits, it often needs to target different labels. Classifiers are trained separately, which often takes a lot of time for model training, which affects the recognition efficiency. Therefore, how to provide a multi-tag recognition method that can improve the recognition accuracy and recognition efficiency of user portrait tags has become an urgent technical problem to be solved.
发明内容Contents of the invention
本申请实施例的主要目的在于提出一种多标签识别方法、装置、电子设备及存储介质,旨在提高用户画像标签的识别准确性及识别效率。The main purpose of the embodiment of the present application is to provide a multi-tag identification method, device, electronic device and storage medium, aiming at improving the recognition accuracy and recognition efficiency of user portrait tags.
技术解决方案technical solution
第一方面,本申请实施例提出了一种多标签识别方法,所述方法包括:In the first aspect, the embodiment of the present application proposes a multi-label identification method, the method comprising:
获取原始数据,其中,所述原始数据包括用户基础数据、用户行为数据以及用户评论数据;Obtaining raw data, wherein the raw data includes user basic data, user behavior data and user comment data;
对所述用户基础数据进行归一化处理,得到用户基础特征;performing normalization processing on the user basic data to obtain user basic features;
通过预先训练的图卷积模型对所述用户行为数据进行特征提取,得到行为特征矩阵;performing feature extraction on the user behavior data through a pre-trained graph convolution model to obtain a behavior feature matrix;
对所述用户评论数据进行分词处理,得到评论文本词段向量;Carry out word segmentation processing to described user comment data, obtain comment text word segment vector;
将所述评论文本词段向量输入至预先训练的对比学习模型中,以使所述评论文本词段向量与所述对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量;The comment text word segment vector is input into the pre-trained comparative learning model, so that the comment text word segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied to obtain the comment word embedding vector;
对所述用户基础特征、所述行为特征矩阵以及所述评论词嵌入向量进行融合处理,得到标准画像特征向量;Perform fusion processing on the user basic features, the behavior feature matrix and the comment word embedding vector to obtain a standard portrait feature vector;
通过预先训练的标签识别模型对所述标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值;Carrying out label recognition processing on the standard portrait feature vector through a pre-trained label recognition model to obtain the probability value of each preset portrait label;
根据所述概率值与预设概率阈值的大小关系,得到目标画像标签。According to the magnitude relationship between the probability value and the preset probability threshold, the target portrait label is obtained.
第二方面,本申请实施例提出了一种多标签识别装置,所述装置包括:In the second aspect, the embodiment of the present application proposes a multi-tag identification device, which includes:
数据获取模块,用于获取原始数据,其中,所述原始数据包括用户基础数据、用户行为数据以及用户评论数据;A data acquisition module, configured to acquire raw data, wherein the raw data includes user basic data, user behavior data and user comment data;
归一化模块,用于对所述用户基础数据进行归一化处理,得到用户基础特征;A normalization module, configured to perform normalization processing on the user basic data to obtain user basic features;
特征提取模块,用于通过预先训练的图卷积模型对所述用户行为数据进行特征提取,得到行为特征矩阵;Feature extraction module, for carrying out feature extraction to described user behavior data by pre-trained graph convolution model, obtains behavior feature matrix;
分词模块,用于对所述用户评论数据进行分词处理,得到评论文本词段向量;A word segmentation module, configured to perform word segmentation processing on the user comment data to obtain a comment text word segment vector;
对比学习模块,用于将所述评论文本词段向量输入至预先训练的对比学习模型中,以使 所述评论文本词段向量与所述对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量;A comparative learning module, for inputting the comment text word segment vector into a pre-trained comparative learning model, so that the comment text word segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied, Get the comment embedding vector;
融合模块,用于对所述用户基础特征、所述行为特征矩阵以及所述评论词嵌入向量进行融合处理,得到标准画像特征向量;The fusion module is used to perform fusion processing on the user basic features, the behavior feature matrix and the comment word embedding vector to obtain a standard portrait feature vector;
标签识别模块,用于通过预先训练的标签识别模型对所述标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值;A tag recognition module, configured to perform tag recognition processing on the standard portrait feature vector through a pre-trained tag recognition model to obtain the probability value of each preset portrait tag;
比较模块,用于根据所述概率值与预设概率阈值的大小关系,得到目标画像标签。The comparison module is used to obtain the target portrait label according to the magnitude relationship between the probability value and the preset probability threshold.
第三方面,本申请实施例提出了一种电子设备,所述电子设备包括存储器、处理器、存储在所述存储器上并可在所述处理器上运行的程序以及用于实现所述处理器和所述存储器之间的连接通信的数据总线,所述程序被所述处理器执行时实现一种多标签识别方法,其中,所述多标签识别方法包括:获取原始数据,其中,所述原始数据包括用户基础数据、用户行为数据以及用户评论数据;对所述用户基础数据进行归一化处理,得到用户基础特征;通过预先训练的图卷积模型对所述用户行为数据进行特征提取,得到行为特征矩阵;对所述用户评论数据进行分词处理,得到评论文本词段向量;将所述评论文本词段向量输入至预先训练的对比学习模型中,以使所述评论文本词段向量与所述对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量;对所述用户基础特征、所述行为特征矩阵以及所述评论词嵌入向量进行融合处理,得到标准画像特征向量;通过预先训练的标签识别模型对所述标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值;根据所述概率值与预设概率阈值的大小关系,得到目标画像标签。In the third aspect, the embodiment of the present application provides an electronic device, the electronic device includes a memory, a processor, a program stored in the memory and operable on the processor, and a program for implementing the processor A data bus connecting and communicating with the memory, when the program is executed by the processor, a multi-label identification method is implemented, wherein the multi-label identification method includes: obtaining original data, wherein the original The data includes user basic data, user behavior data, and user comment data; the user basic data is normalized to obtain user basic features; the user behavior data is extracted through a pre-trained graph convolution model to obtain Behavior feature matrix; word segmentation processing is carried out to described user comment data, obtain comment text phrase vector; Described comment text phrase vector is input in the comparative learning model of pre-training, so that described comment text phrase vector and all The reference word embedding matrix in the comparative learning model is multiplied by matrix to obtain the comment word embedding vector; the user basic feature, the behavior feature matrix and the comment word embedding vector are fused to obtain a standard portrait feature vector; Perform label recognition processing on the standard portrait feature vector through a pre-trained label recognition model to obtain the probability value of each preset portrait label; obtain the target portrait label according to the magnitude relationship between the probability value and the preset probability threshold.
第四方面,本申请实施例提出了一种存储介质,所述存储介质为计算机可读存储介质,用于计算机可读存储,所述存储介质存储有一个或者多个程序,所述一个或者多个程序可被一个或者多个处理器执行,以实现一种多标签识别方法,其中,所述多标签识别方法包括:获取原始数据,其中,所述原始数据包括用户基础数据、用户行为数据以及用户评论数据;对所述用户基础数据进行归一化处理,得到用户基础特征;通过预先训练的图卷积模型对所述用户行为数据进行特征提取,得到行为特征矩阵;对所述用户评论数据进行分词处理,得到评论文本词段向量;将所述评论文本词段向量输入至预先训练的对比学习模型中,以使所述评论文本词段向量与所述对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量;对所述用户基础特征、所述行为特征矩阵以及所述评论词嵌入向量进行融合处理,得到标准画像特征向量;通过预先训练的标签识别模型对所述标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值;根据所述概率值与预设概率阈值的大小关系,得到目标画像标签。In a fourth aspect, the embodiment of the present application provides a storage medium, the storage medium is a computer-readable storage medium for computer-readable storage, the storage medium stores one or more programs, and the one or more This program can be executed by one or more processors to implement a multi-label identification method, wherein the multi-label identification method includes: obtaining original data, wherein the original data includes user basic data, user behavior data, and User comment data; normalize the user basic data to obtain user basic features; perform feature extraction on the user behavior data through a pre-trained graph convolution model to obtain a behavior feature matrix; Carry out participle processing, obtain comment text phrase vector; Described comment text phrase vector is input in the comparative learning model of pre-training, so that described comment text phrase vector and the reference word embedding matrix in described contrast learning model Perform matrix multiplication to obtain comment word embedding vectors; perform fusion processing on the user basic features, the behavior feature matrix, and the comment word embedding vectors to obtain standard portrait feature vectors; The standard portrait feature vector performs label recognition processing to obtain the probability value of each preset portrait label; according to the magnitude relationship between the probability value and the preset probability threshold, the target portrait label is obtained.
有益效果Beneficial effect
本申请提出的多标签识别方法、装置、电子设备及存储介质,能够极大地缩短模型训练时间,提高了识别效率,能够提高用户画像标签的识别准确性。The multi-tag recognition method, device, electronic equipment, and storage medium proposed in this application can greatly shorten the model training time, improve the recognition efficiency, and improve the recognition accuracy of user portrait tags.
附图说明Description of drawings
附图用来提供对本申请技术方案的进一步理解,并且构成说明书的一部分,与本申请的实施例一起用于解释本申请的技术方案,并不构成对本申请技术方案的限制。The accompanying drawings are used to provide a further understanding of the technical solution of the present application, and constitute a part of the specification, and are used together with the embodiments of the present application to explain the technical solution of the present application, and do not constitute a limitation to the technical solution of the present application.
图1是本申请实施例提供的多标签识别方法的流程图;Fig. 1 is a flow chart of the multi-tag identification method provided by the embodiment of the present application;
图2是图1中的步骤S103的流程图;Fig. 2 is the flowchart of step S103 in Fig. 1;
图3是图1中的步骤S104的流程图;Fig. 3 is the flowchart of step S104 in Fig. 1;
图4是图1中的步骤S105的流程图;Fig. 4 is the flowchart of step S105 in Fig. 1;
图5是本申请实施例提供的多标签识别方法的另一流程图;Fig. 5 is another flow chart of the multi-tag identification method provided by the embodiment of the present application;
图6是图1中的步骤S107的流程图;Fig. 6 is the flowchart of step S107 in Fig. 1;
图7是图1中的步骤S108的流程图;Fig. 7 is the flowchart of step S108 in Fig. 1;
图8是本申请实施例提供的多标签识别装置的结构示意图;Fig. 8 is a schematic structural diagram of a multi-tag identification device provided by an embodiment of the present application;
图9是本申请实施例提供的电子设备的硬件结构示意图。FIG. 9 is a schematic diagram of a hardware structure of an electronic device provided by an embodiment of the present application.
本发明的实施方式Embodiments of the present invention
为了使本申请的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处所描述的具体实施例仅用以解释本申请,并不用于限定本申请。In order to make the purpose, technical solution and advantages of the present application clearer, the present application will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present application, not to limit the present application.
需要说明的是,虽然在装置示意图中进行了功能模块划分,在流程图中示出了逻辑顺序,但是在某些情况下,可以以不同于装置中的模块划分,或流程图中的顺序执行所示出或描述的步骤。说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。It should be noted that although the functional modules are divided in the schematic diagram of the device, and the logical sequence is shown in the flowchart, in some cases, it can be executed in a different order than the module division in the device or the flowchart in the flowchart. steps shown or described. The terms "first", "second" and the like in the specification and claims and the above drawings are used to distinguish similar objects, and not necessarily used to describe a specific sequence or sequence.
除非另有定义,本文所使用的所有的技术和科学术语与属于本申请的技术领域的技术人员通常理解的含义相同。本文中所使用的术语只是为了描述本申请实施例的目的,不是旨在限制本申请。Unless otherwise defined, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the technical field to which this application belongs. The terms used herein are only for the purpose of describing the embodiments of the present application, and are not intended to limit the present application.
目前,在对互联网用户进行画像标签时,常常采用人工标注或者机器学习的方式来对画像标签进行识别和分类。当采用人工标注方式时,往往需要经过长时间的标记处理,且出错率较高,影响识别准确性;而当采用机器学习的方式对多标签画像进行识别时,往往需要针对不同的标签类别,分别训练分类器,往往需要花费较多的时间进行模型训练,影响识别效率。因此,如何提供一种多标签识别方法,能够提高用户画像标签的识别准确性及识别效率,成为了亟待解决的技术问题。At present, when labeling portraits of Internet users, manual labeling or machine learning is often used to identify and classify the portrait labels. When manual labeling is used, it often takes a long time to label, and the error rate is high, which affects the accuracy of recognition; and when machine learning is used to identify multi-label portraits, it is often necessary to target different label categories. Training the classifiers separately often takes a lot of time for model training, which affects the recognition efficiency. Therefore, how to provide a multi-tag recognition method that can improve the recognition accuracy and recognition efficiency of user portrait tags has become an urgent technical problem to be solved.
基于此,本申请实施例提供了一种多标签识别方法、装置、电子设备及存储介质,旨在提高用户画像标签的识别准确性。Based on this, embodiments of the present application provide a multi-tag identification method, device, electronic equipment, and storage medium, aiming at improving the identification accuracy of user portrait tags.
本申请实施例提供的多标签识别方法、装置、电子设备及存储介质,具体通过如下实施例进行说明,首先描述本申请实施例中的多标签识别方法。The multi-tag identification method, device, electronic device, and storage medium provided in the embodiments of the present application are specifically described through the following embodiments. First, the multi-tag identification method in the embodiment of the present application is described.
本申请实施例可以基于人工智能技术对相关的数据进行获取和处理。其中,人工智能(Artificial Intelligence,AI)是利用数字计算机或者数字计算机控制的机器模拟、延伸和扩展人的智能,感知环境、获取知识并使用知识获得最佳结果的理论、方法、技术及应用系统。The embodiments of the present application may acquire and process relevant data based on artificial intelligence technology. Among them, artificial intelligence (AI) is the theory, method, technology and application system that uses digital computers or machines controlled by digital computers to simulate, extend and expand human intelligence, perceive the environment, acquire knowledge and use knowledge to obtain the best results. .
人工智能基础技术一般包括如传感器、专用人工智能芯片、云计算、分布式存储、大数据处理技术、操作/交互系统、机电一体化等技术。人工智能软件技术主要包括计算机视觉技术、机器人技术、生物识别技术、语音处理技术、自然语言处理技术以及机器学习/深度学习等几大方向。Artificial intelligence basic technologies generally include technologies such as sensors, dedicated artificial intelligence chips, cloud computing, distributed storage, big data processing technology, operation/interaction systems, and mechatronics. Artificial intelligence software technology mainly includes computer vision technology, robotics technology, biometrics technology, speech processing technology, natural language processing technology, and machine learning/deep learning.
本申请实施例提供的多标签识别方法,涉及人工智能及数字医疗技术领域。本申请实施例提供的多标签识别方法可应用于终端中,也可应用于服务器端中,还可以是运行于终端或服务器端中的软件。在一些实施例中,终端可以是智能手机、平板电脑、笔记本电脑、台式计算机等;服务器端可以配置成独立的物理服务器,也可以配置成多个物理服务器构成的服务器集群或者分布式系统,还可以配置成提供云服务、云数据库、云计算、云函数、云存储、网络服务、云通信、中间件服务、域名服务、安全服务、CDN以及大数据和人工智能平台等基础云计算服务的云服务器;软件可以是实现多标签识别方法的应用等,但并不局限于以上形式。The multi-tag identification method provided in the embodiment of the present application relates to the fields of artificial intelligence and digital medical technology. The multi-tag identification method provided in the embodiment of the present application can be applied to a terminal, can also be applied to a server, and can also be software running on a terminal or a server. In some embodiments, the terminal can be a smart phone, a tablet computer, a notebook computer, a desktop computer, etc.; the server end can be configured as an independent physical server, or can be configured as a server cluster or a distributed system composed of multiple physical servers, or It can be configured as a cloud that provides basic cloud computing services such as cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communications, middleware services, domain name services, security services, CDN, and big data and artificial intelligence platforms. The server; the software may be an application to realize the multi-label identification method, but is not limited to the above forms.
本申请可用于众多通用或专用的计算机系统环境或配置中。例如:个人计算机、服务器计算机、手持设备或便携式设备、平板型设备、多处理器系统、基于微处理器的系统、置顶盒、可编程的消费电子设备、网络PC、小型计算机、大型计算机、包括以上任何系统或设备的分布式计算环境等等。本申请可以在由计算机执行的计算机可执行指令的一般上下文中描述,例如程序模块。一般地,程序模块包括执行特定任务或实现特定抽象数据类型的例程、程序、对象、组件、数据结构等等。也可以在分布式计算环境中实践本申请,在这些分布式 计算环境中,由通过通信网络而被连接的远程处理设备来执行任务。在分布式计算环境中,程序模块可以位于包括存储设备在内的本地和远程计算机存储介质中。The application can be used in numerous general purpose or special purpose computer system environments or configurations. Examples: personal computers, server computers, handheld or portable devices, tablet-type devices, multiprocessor systems, microprocessor-based systems, set-top boxes, programmable consumer electronics, network PCs, minicomputers, mainframe computers, including A distributed computing environment for any of the above systems or devices, etc. This application may be described in the general context of computer-executable instructions, such as program modules, being executed by a computer. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types. The application may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote computer storage media including storage devices.
图1是本申请实施例提供的多标签识别方法的一个可选的流程图,图1中的方法可以包括但不限于包括步骤S101至步骤S108。Fig. 1 is an optional flow chart of the multi-tag identification method provided by the embodiment of the present application. The method in Fig. 1 may include but not limited to steps S101 to S108.
步骤S101,获取原始数据,其中,原始数据包括用户基础数据、用户行为数据以及用户评论数据;Step S101, obtaining original data, wherein the original data includes user basic data, user behavior data and user comment data;
步骤S102,对用户基础数据进行归一化处理,得到用户基础特征;Step S102, performing normalization processing on the basic user data to obtain basic user features;
步骤S103,通过预先训练的图卷积模型对用户行为数据进行特征提取,得到行为特征矩阵;Step S103, using the pre-trained graph convolution model to perform feature extraction on user behavior data to obtain a behavior feature matrix;
步骤S104,对用户评论数据进行分词处理,得到评论文本词段向量;Step S104, performing word segmentation processing on user comment data to obtain comment text word segment vectors;
步骤S105,将评论文本词段向量输入至预先训练的对比学习模型中,以使评论文本词段向量与对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量;Step S105, inputting the comment text word segment vector into the pre-trained comparative learning model, so that the comment text word segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied to obtain the comment word embedding vector;
步骤S106,对用户基础特征、行为特征矩阵以及评论词嵌入向量进行融合处理,得到标准画像特征向量;Step S106, performing fusion processing on user basic features, behavior feature matrix and comment word embedding vectors to obtain standard portrait feature vectors;
步骤S107,通过预先训练的标签识别模型对标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值;Step S107, using the pre-trained label recognition model to perform label recognition processing on the standard portrait feature vector to obtain the probability value of each preset portrait label;
步骤S108,根据概率值与预设概率阈值的大小关系,得到目标画像标签。Step S108, according to the magnitude relationship between the probability value and the preset probability threshold, the target portrait label is obtained.
经过以上步骤S101至步骤S108,本申请的多标签识别方法通过一个标签识别模型能够实现对不同画像标签的识别,相较于传统技术中需要针对不同标签类别分别训练分类器,能够极大地缩短模型训练时间,提高了识别效率。同时,本申请的多标签识别方法对于不同类型的用户数据分别进行了相应的数据预处理,使得获取到的标准画像特征向量能够更加符合识别需求,能够提高用户画像标签的识别准确性。After the above steps S101 to S108, the multi-label recognition method of the present application can realize the recognition of different portrait labels through a label recognition model. Compared with the traditional technology, which needs to train classifiers for different label categories separately, the model can be greatly shortened. The training time improves the recognition efficiency. At the same time, the multi-label recognition method of the present application performs corresponding data preprocessing for different types of user data, so that the obtained standard portrait feature vectors can better meet the recognition requirements, and can improve the recognition accuracy of user portrait labels.
在一些实施例中,在执行步骤S101时,可以通过网络爬虫的方式,从预设的多个数据源爬取原始用户数据,其中基础数据包括用户的性别、学历和年龄段等等;行为数据包括用户在课程内容展示的点击数据以及课程页面内推荐课程的点击数据等等;评论数据为用户对课程的文本评论数据等等。In some embodiments, when step S101 is executed, the original user data can be crawled from multiple preset data sources by means of a web crawler, wherein the basic data includes the user's gender, education background, age group, etc.; behavior data Including the user's click data on the display of the course content and the click data of the recommended courses on the course page, etc.; the comment data is the user's text comment data on the course, etc.
在一些实施例中,在执行步骤S102时,可以根据预设的归一化条件,对不同类型的基础数据设置有一系列的数字,例如,基础数据包括性别、学历、年龄段,把性别分为{0,1}的集合,0代表女性,1代表男性。学历分为{1,2,3,4,5,6,7,8}的集合,1代表小学,2代表初中,3代表中专,4代表高中,5代表大专,6代表本科,7代表硕士,8代表博士。年龄段分为{5,6,7,8,9,0}的集合,5代表50后,6代表60后,依次类推。In some embodiments, when step S102 is executed, a series of numbers can be set for different types of basic data according to preset normalization conditions. A collection of {0,1}, 0 for female and 1 for male. Educational background is divided into {1,2,3,4,5,6,7,8} set, 1 represents primary school, 2 represents junior high school, 3 represents technical secondary school, 4 represents high school, 5 represents junior college, 6 represents undergraduate, 7 represents Master, 8 represents Ph.D. The age group is divided into {5, 6, 7, 8, 9, 0} sets, 5 represents the post-50s, 6 represents the post-60s, and so on.
请参阅图2,在一些实施例中,步骤S103可以包括但不限于包括步骤S201至步骤S204:Referring to FIG. 2, in some embodiments, step S103 may include but not limited to include steps S201 to S204:
步骤S201,将用户行为数据映射到预设的向量空间,得到用户行为特征向量;Step S201, mapping user behavior data to a preset vector space to obtain user behavior feature vectors;
步骤S202,根据预设的课程类型和用户行为特征向量,构建行为特征图;Step S202, constructing a behavior feature map according to the preset course types and user behavior feature vectors;
步骤S203,对行为特征图进行图卷积处理,得到行为度矩阵和行为邻接矩阵;Step S203, performing graph convolution processing on the behavior feature map to obtain a behavior degree matrix and a behavior adjacency matrix;
步骤S204,对行为度矩阵和行为矩阵进行做差处理,得到行为特征矩阵。Step S204, performing difference processing on the behavior degree matrix and the behavior matrix to obtain a behavior characteristic matrix.
具体地,在步骤S201中,可以采用MLP网络对用户行为数据行语义空间到向量空间上的映射处理,将用户行为数据映射到预先设定的向量空间中,得到用户行为特征向量。Specifically, in step S201, the user behavior data can be mapped to a preset vector space by using the MLP network to map the semantic space of the user behavior data to the vector space, and obtain the user behavior feature vector.
在步骤S202中,将每一预设的课程记为一个节点,对用户的行为数据进行分析,若检测到用户通过一个课程页面的推荐模块点击到另一课程,则建立这两个课程之间的边。根据这种映射关系,对每一课程类型和用户行为特征向量进行关系构建,得到一个无向图,该无向图即为行为特征图。In step S202, record each preset course as a node, analyze the user's behavior data, and if it is detected that the user clicks on another course through the recommendation module of a course page, then establish a link between the two courses. side. According to this mapping relationship, the relationship between each course type and user behavior feature vector is constructed to obtain an undirected graph, which is the behavior feature graph.
在步骤S203中,行为特征图可以表示为G=(V,E),V代表节点,E代表边。该行为特征图的拉普拉斯矩阵可以定义为L=D-A,L为拉普拉斯矩阵,D为对角度矩阵(由于对角线上的元素是顶点的度,即对角度矩阵是指元素链接的元素个数);A为邻接矩阵,表示任意两个顶点之间的邻接关系,若两个顶点之间邻接,则邻接矩阵为1,若两个顶点之间不邻接,则邻 接矩阵为0;因而,通过对行为特征图进行图卷积处理,能够实现对行为特征图的拉普拉斯变换,得到行为度矩阵(即对角度矩阵D)和行为邻接矩阵(即邻接矩阵A)。In step S203, the behavior feature graph can be expressed as G=(V, E), where V represents a node and E represents an edge. The Laplacian matrix of this behavior feature map can be defined as L=D-A, L is the Laplacian matrix, D is the angle matrix (because the element on the diagonal is the degree of the vertex, the angle matrix refers to the element The number of linked elements); A is an adjacency matrix, which represents the adjacency relationship between any two vertices. If the two vertices are adjacent, the adjacency matrix is 1. If the two vertices are not adjacent, the adjacency matrix is 0; therefore, by performing graph convolution processing on the behavior feature map, the Laplace transform of the behavior feature map can be realized, and the behavior degree matrix (ie, the angle matrix D) and the behavior adjacency matrix (ie, the adjacency matrix A) can be obtained.
在步骤S204中,由于拉普拉斯矩阵与图的性质满足L=D-A,即对行为度矩阵D和行为邻接矩阵A进行做差处理,可以得到行为特征矩阵L1。In step S204, since the property of the Laplacian matrix and the graph satisfies L=D-A, that is, the behavior degree matrix D and the behavior adjacency matrix A are subtracted to obtain the behavior characteristic matrix L1.
需要说明的是,该图卷积模型的图卷积层可以表示为公式(1)所示:It should be noted that the graph convolution layer of this graph convolution model can be expressed as shown in formula (1):
Figure PCTCN2022090726-appb-000001
Figure PCTCN2022090726-appb-000001
其中,y为输出值,σ为sigmoid激活函数。L是拉普拉斯矩阵,x是输入的标注行为特征图,j是拉普拉斯矩阵的行数,其中,j在一般情况下远小于行为特征图中的节点的数量。α为权重矩阵,权重矩阵的参数值在图卷积模型初始化的时候随机生成,后期可以通过对图卷积模型的训练来调整,具体地,计算标注行为特征与预测特征之间的误差,再把误差进行反向传播来更新参数值,以实现对图卷积模型的优化。Among them, y is the output value and σ is the sigmoid activation function. L is the Laplacian matrix, x is the input labeled behavioral feature map, and j is the number of rows of the Laplacian matrix, where j is generally much smaller than the number of nodes in the behavioral feature map. α is the weight matrix. The parameter values of the weight matrix are randomly generated when the graph convolution model is initialized, and can be adjusted later by training the graph convolution model. Specifically, calculate the error between the marked behavior features and the predicted features, and then The error is backpropagated to update the parameter value to optimize the graph convolution model.
请参阅图3,在一些实施例中,步骤S104可以包括但不限于包括步骤S301至步骤S302:Referring to FIG. 3, in some embodiments, step S104 may include but not limited to include steps S301 to S302:
步骤S301,通过预设的分词器对用户评论数据进行分词处理,得到评论文本词段;Step S301, performing word segmentation processing on the user comment data through a preset word segmenter to obtain comment text segments;
步骤S302,对评论文本词段进行编码处理,得到评论文本词段向量。Step S302, encoding the word segment of the review text to obtain a word segment vector of the review text.
具体地,在步骤S301中,在利用Jieba分词器对用户评论数据进行分词处理时,首先通过对照Jieba分词器内的词典生成该用户评论数据对应的有向无环图,再根据预设的选择模式和词典寻找有向无环图上的最短路径,根据最短路径对该用户评论数据进行截取,或者直接对该用户评论数据进行截取,得到评论文本词段。Specifically, in step S301, when using the Jieba tokenizer to perform word segmentation processing on the user comment data, first generate a directed acyclic graph corresponding to the user comment data by comparing the dictionary in the Jieba tokenizer, and then according to the preset selection Patterns and dictionaries search for the shortest path on the directed acyclic graph, and intercept the user comment data according to the shortest path, or directly intercept the user comment data to obtain comment text segments.
进一步地,对于不在词典中的评论文本词段,可以使用HMM(隐马尔科夫模型)进行新词发现。具体地,将字符在评论文本词段中的位置B、M、E、S作为隐藏状态,字符是观测状态,其中,B/M/E/S分别代表出现在词头、词中、词尾以及单字成词。使用词典文件分别存储字符之间的表现概率矩阵、初始概率向量和转移概率矩阵。再利用维特比算法对最大可能的隐藏状态进行求解,从而得到评论文本词段。Further, HMM (Hidden Markov Model) can be used to discover new words for comment text segments that are not in the dictionary. Specifically, the positions B, M, E, and S of the characters in the comment text segment are regarded as hidden states, and the characters are observed states, where B/M/E/S represent the words that appear at the beginning, middle, end, and word respectively into words. A dictionary file is used to store representation probability matrix, initial probability vector and transition probability matrix between characters respectively. Then use the Viterbi algorithm to solve the maximum possible hidden state, so as to obtain the comment text segment.
在步骤S302中,可以是利用预设的BERT编码器对评论文本词段进行编码处理,使得评论文本词段上的每个字符都带上对应的编码,从而得到评论文本词段向量。In step S302, a preset BERT encoder may be used to encode the word segment of the comment text, so that each character on the word segment of the comment text carries a corresponding code, thereby obtaining a word segment vector of the comment text.
在一些实施例中,在步骤S105之前,该方法还包括预先训练对比学习模型,具体可以包括但不限于包括步骤a至步骤f:In some embodiments, before step S105, the method further includes pre-training a contrastive learning model, which may specifically include but not limited to steps a to f:
a、获取样本用户数据;a. Obtain sample user data;
b、通过对比学习模型对样本用户数据进行映射处理和编码处理,得到初始嵌入数据;b. Mapping and encoding the sample user data through the comparative learning model to obtain the initial embedded data;
c、根据初始嵌入数据构建样本对,其中,样本对包括正例对和负例对;c. Construct sample pairs according to the initial embedding data, wherein the sample pairs include positive example pairs and negative example pairs;
d、将样本对输入到对比学习模型中;d. Input the sample pair into the contrastive learning model;
e、通过对比学习模型的损失函数计算出正例对的第一相似度和负例对的第二相似度。e. Calculate the first similarity degree of the positive example pair and the second similarity degree of the negative example pair by comparing the loss function of the learning model.
f、根据第一相似度和第二相似度对对比学习模型的损失函数进行优化,以更新对比学习模型。f. Optimizing the loss function of the contrastive learning model according to the first similarity and the second similarity, so as to update the contrastive learning model.
具体地,执行步骤a和步骤b,首先获取样本用户数据,对样本用户数据进行编码处理,将样本用户数据映射至嵌入空间、并对样本用户数据进行向量表示,从而可以得到初始嵌入数据(即初始embedding数据),该初始嵌入数据包括正样本数据和负样本数据。Specifically, to execute step a and step b, first obtain the sample user data, encode the sample user data, map the sample user data to the embedding space, and perform vector representation on the sample user data, so as to obtain the initial embedded data (ie Initial embedding data), the initial embedding data includes positive sample data and negative sample data.
在一些实施例的步骤c中,通过dropout mask机制对初始嵌入数据进行数据增强处理;本申请实施例通过dropout mask机制替换了传统的数据增强方法,即将同一个样本数据两次输入dropout编码器得到的两个向量作为对比学习的正例对,效果就足够好了,因为比如BERT内部每次dropout都随机会生成一个不同的dropout mask,所以只需要将同一个样本数据(即本实施例的初始嵌入数据)输入至simCSE模型两次,得到的两个向量就是应用两次不同dropout mask的结果了。可以理解的是,dropout mask是一种网络模型的随机,是对模型参数W的mask,起到防止过拟合的作用。In step c of some embodiments, the data enhancement processing is performed on the initial embedded data through the dropout mask mechanism; the embodiment of the present application replaces the traditional data enhancement method through the dropout mask mechanism, that is, the same sample data is input into the dropout encoder twice to obtain The two vectors of the two vectors are used as positive example pairs for comparative learning, and the effect is good enough, because for example, a different dropout mask is randomly generated for each dropout inside BERT, so only the same sample data (that is, the initial Embedding data) is input to the simCSE model twice, and the two vectors obtained are the results of applying two different dropout masks. It is understandable that the dropout mask is a random network model, which is the mask of the model parameter W, which prevents overfitting.
在一个batch中,经过数据增强处理得到的数据(即第一向量和第二向量)是正例对,未经过数据增强的其他数据为负例对。本申请实施例中,可以将一个batch中的其中一部分初始嵌入数据经过数据增强处理得到正例对,另一部分初始嵌入数据作为负例对。In a batch, the data (that is, the first vector and the second vector) obtained through data enhancement processing are positive example pairs, and other data that have not undergone data enhancement are negative example pairs. In the embodiment of the present application, some of the initial embedded data in a batch can be processed through data enhancement to obtain positive example pairs, and the other part of the initial embedded data can be used as negative example pairs.
进一步地,执行步骤d,将样本对输入到对比学习模型中。Further, step d is performed to input the sample pair into the contrastive learning model.
在一些实施例的步骤e中,第一相似度和第二相似度均为余弦相似度。In step e of some embodiments, both the first similarity and the second similarity are cosine similarity.
在一些实施例中,步骤f可以包括但不限于包括:In some embodiments, step f may include, but is not limited to include:
将第一相似度最大化为第一数值和将第二相似度最小化为第一数值,以对损失函数进行优化;其中,第一相似度为损失函数的分子,第一相似度和第二相似度为损失函数的分母,第一数值取值为1,第二数值取值为0。该损失函数中,分子是对应正例对的第一相似度,分母是第一相似度以及所有负例对的第二相似度,然后将分子和分母构成的分子式的值包装在-log()中,这样最大化分子且最小化分母,就能实现最小化损失函数。本申请实施例中,最小化损失函数infoNCE loss,就是最大化分子且最小化分母,也就是最大化正例对的第一相似度且最小化负例对的第二相似度,并对该损失函数进行最小化,实现对损失函数的优化。更具体地,损失函数为公式(2)所示:Maximize the first similarity to the first value and minimize the second similarity to the first value to optimize the loss function; where the first similarity is the numerator of the loss function, the first similarity and the second The similarity is the denominator of the loss function, the first value is 1, and the second value is 0. In this loss function, the numerator is the first similarity of the corresponding positive example pair, the denominator is the first similarity and the second similarity of all negative example pairs, and then the value of the molecular formula composed of the numerator and the denominator is wrapped in -log() In this way, the loss function can be minimized by maximizing the numerator and minimizing the denominator. In the embodiment of this application, minimizing the loss function infoNCE loss is to maximize the numerator and minimize the denominator, that is, to maximize the first similarity of the positive pair and minimize the second similarity of the negative pair, and the loss The function is minimized to realize the optimization of the loss function. More specifically, the loss function is shown in formula (2):
Figure PCTCN2022090726-appb-000002
Figure PCTCN2022090726-appb-000002
该损失函数中,l i为损失函数的损失值,正例对是<z,z′>,N是batch的大小(N是变量),该损失函数表示的是第i个样本要与batch中的每个样本计算相似度,batch里的每个样本都会按照该损失函数进行计算,因此,该损失函数表示的是样本i的损失(loss);该损失函数中,分子是正例对的相似度,分母是正例对以及所有负例对的相似度,然后将该值包装在-log()中,这样最大化分子且最小化分母,就能实现最小化损失函数。 In this loss function, l i is the loss value of the loss function, the positive example pair is <z, z′>, N is the size of the batch (N is a variable), and the loss function indicates that the i-th sample should be compared with the batch The similarity is calculated for each sample of , and each sample in the batch will be calculated according to the loss function. Therefore, the loss function represents the loss of sample i; in the loss function, the numerator is the similarity of the positive pair , the denominator is the similarity between positive pairs and all negative pairs, and then wrap this value in -log(), so that maximizing the numerator and minimizing the denominator can minimize the loss function.
请参阅图4,在一些实施例中,步骤S105可以包括但不限于包括步骤S401至步骤S402:Referring to FIG. 4, in some embodiments, step S105 may include but not limited to include steps S401 to S402:
步骤S401,将评论文本词段向量输入到对比学习模型中,以使评论文本词段向量与参考词嵌入矩阵进行矩阵相乘,得到多个基本词嵌入向量;Step S401, input the comment text word segment vector into the comparative learning model, so that the comment text word segment vector and the reference word embedding matrix are matrix multiplied to obtain a plurality of basic word embedding vectors;
步骤S402,对基本词嵌入向量进行映射处理,得到评论词嵌入向量。Step S402, performing mapping processing on the basic word embedding vectors to obtain review word embedding vectors.
具体地,执行步骤S401,通过训练对比模型可以使得对比模型内的参考词嵌入矩阵的数值将被完全固定下来,对比模型的其他模型参数也被固定。因而,将评论文本词段向量输入到对比模型中,可以利用固定的参考词嵌入矩阵与每一评论文本词段向量进行矩阵相乘,得到多个基本词嵌入向量。Specifically, step S401 is executed, and the value of the reference word embedding matrix in the comparison model can be completely fixed by training the comparison model, and other model parameters of the comparison model are also fixed. Therefore, inputting the comment text word segment vector into the comparison model, the fixed reference word embedding matrix can be used to perform matrix multiplication with each comment text word segment vector to obtain multiple basic word embedding vectors.
在步骤S402中,利用对比模型中固定的MLP网络对基本词嵌入向量进行映射处理,得到评论词嵌入向量。其中,MLP网络包括linear层、ReLu激活函数以及linear层。In step S402, use the fixed MLP network in the comparison model to perform mapping processing on the basic word embedding vector to obtain the comment word embedding vector. Among them, the MLP network includes a linear layer, a ReLu activation function, and a linear layer.
在一些实施例中,在执行步骤S106时,分别对基础特征数据、行为特征矩阵进行向量化处理,得到基础特征向量,行为特征向量,进而对基础特征向量,行为特征向量和词嵌入特征向量进行融合处理,得到标准特征向量。例如,标准特征向量X=[性别,学历,年龄段,[GCN],[BERT]]。其中GCN为256维的向量,BERT为512维的向量,X为3+256+512的向量。In some embodiments, when step S106 is executed, the basic feature data and the behavioral feature matrix are respectively vectorized to obtain the basic feature vector and the behavioral feature vector, and then the basic feature vector, the behavioral feature vector and the word embedding feature vector are performed Fusion processing to get the standard feature vector. For example, the standard feature vector X=[gender, education, age group, [GCN], [BERT]]. Among them, GCN is a 256-dimensional vector, BERT is a 512-dimensional vector, and X is a 3+256+512 vector.
请参阅图5,在一些实施例中,在步骤S107之前,该方法还包括预先训练标签识别模型,具体可以包括但不限于包括步骤S501至步骤S505:Please refer to FIG. 5 , in some embodiments, before step S107, the method further includes pre-training the label recognition model, which may specifically include but not limited to steps S501 to S505:
步骤S501,获取标注用户数据;Step S501, acquiring marked user data;
步骤S502,对标注用户数据进行特征提取,得到样本特征向量;Step S502, performing feature extraction on the labeled user data to obtain sample feature vectors;
步骤S503,将样本特征向量输入到标签识别模型中;Step S503, inputting the sample feature vector into the label recognition model;
步骤S504,通过标签识别模型的损失函数计算出每一画像标签类别的样本概率预测值;Step S504, calculate the sample probability prediction value of each portrait label category through the loss function of the label recognition model;
步骤S505,根据样本概率预测值对标签识别模型的损失函数进行优化,以更新标签识别模型。Step S505, optimize the loss function of the label recognition model according to the sample probability prediction value, so as to update the label recognition model.
需要说明的是,该标签识别模型可以为textcnn模型,该标签识别模型包括Embedding 层,卷积层,池化层和输出层。通常经过标签识别模型的Embedding层可以采用ELMO,GLOVE,Word2Vector,Bert等算法将输入的文本数据生成一个稠密向量。进而通过标签识别模型的卷积层和池化层对该稠密向量进行卷积处理和池化处理,得到目标特征向量,进而将目标特征向量输入至输出层,通过输出层中的预设函数即可对目标特征向量进行分类操作,得到标签特征向量及每一预设类别的概率值大小。It should be noted that the label recognition model can be a textcnn model, and the label recognition model includes an Embedding layer, a convolution layer, a pooling layer and an output layer. Usually, the Embedding layer of the label recognition model can use ELMO, GLOVE, Word2Vector, Bert and other algorithms to generate a dense vector from the input text data. Then, the dense vector is convoluted and pooled through the convolution layer and pooling layer of the label recognition model to obtain the target feature vector, and then the target feature vector is input to the output layer, and the preset function in the output layer is The classification operation can be performed on the target feature vector to obtain the label feature vector and the probability value of each preset category.
首先,执行步骤S501,获取标注用户数据,该标注用户数据包含用户画像类别标签。进而,执行步骤S502,利用MLP网络对标注用户数据进行多次映射处理,得到样本特征向量。Firstly, step S501 is executed to obtain marked user data, where the marked user data includes user portrait category labels. Furthermore, step S502 is executed, using the MLP network to perform multiple mapping processes on the labeled user data to obtain sample feature vectors.
进而,执行步骤S503,将样本特征向量输入到标签识别模型中。Furthermore, step S503 is executed to input the sample feature vector into the label recognition model.
在执行步骤S504时,通过标签识别模型的Embedd ing层将样本特征向量生成一个稠密特征向量,进而通过卷积层和池化层对该稠密特征向量机芯卷积处理和池化处理,得到目标特征向量,进而将目标特征向量输入至输出层,通过损失函数计算出每一画像标签类别的样本概率预测值;其中,损失函数如公式(3)所示:When step S504 is executed, the sample feature vector is generated into a dense feature vector through the Embedding layer of the label recognition model, and then the dense feature vector is convoluted and pooled through the convolution layer and the pooling layer to obtain the target The feature vector, and then input the target feature vector to the output layer, and calculate the sample probability prediction value of each portrait label category through the loss function; where the loss function is shown in formula (3):
Figure PCTCN2022090726-appb-000003
Figure PCTCN2022090726-appb-000003
其中,t为目标值(target),t需要在[0,1]之间取值,由于在本申请实施例中,t作为画像标签类别,因而t取值为0或者1,o表示标签识别模型的概率预测值。Among them, t is the target value (target), and t needs to take a value between [0,1]. Since in the embodiment of this application, t is used as a portrait label category, the value of t is 0 or 1, and o indicates label identification The model's probabilistic predictions.
最后,执行步骤S505,根据样本概率预测值计算标签识别模型的模型损失,即loss值,再利用梯度下降法对loss值进行反向传播,将loss值反馈回标签识别模型,修改标签识别模型的模型参数,重复上述过程,直至loss值满足预设的迭代条件,其中,预设的迭代条件是可以迭代次数达到预设值,或者是损失函数的变化方差小于预设阈值。当loss值满足预设的迭代条件时可以停止反向传播,将最后的模型参数作为最终的模型参数,完成对标签识别模型的更新。Finally, step S505 is executed to calculate the model loss of the label recognition model based on the sample probability prediction value, that is, the loss value, and then use the gradient descent method to backpropagate the loss value, feed the loss value back to the label recognition model, and modify the label recognition model. For model parameters, repeat the above process until the loss value satisfies the preset iteration condition, wherein the preset iteration condition is that the number of iterations can reach the preset value, or the variance of the loss function is smaller than the preset threshold. When the loss value satisfies the preset iteration condition, the backpropagation can be stopped, and the final model parameter can be used as the final model parameter to complete the update of the label recognition model.
请参阅图6,在一些实施例,步骤S107还可以包括但不限于包括步骤S601至步骤S602:Referring to FIG. 6, in some embodiments, step S107 may also include but not limited to steps S601 to S602:
步骤S601,根据预设的标签维度对标准画像特征向量进行重构处理,得到标签特征向量;Step S601, reconstructing the standard portrait feature vector according to the preset label dimension to obtain the label feature vector;
步骤S602,利用预设函数对标签特征向量进行识别处理,得到每一预设画像标签的概率值。Step S602, using a preset function to identify the tag feature vector to obtain the probability value of each preset portrait tag.
具体地,首先执行步骤S601,根据预设的标签维度和编码器对标准画像特征向量进行重构处理,例如,根据自下而上的编码顺序和标签维度,对标准画像特征向量进行编码处理。例如,对标准画像特征向量进行初次编码,得到最底层的标签特征向量z1,然后逐层向上进行下采样处理,得到每一标签维度对应的标签特征向量[z2,z3…,zk]。Specifically, step S601 is first performed to reconstruct the standard image feature vector according to the preset label dimension and encoder, for example, encode the standard image feature vector according to the bottom-up encoding sequence and label dimension. For example, the standard portrait feature vector is encoded for the first time to obtain the bottom label feature vector z1, and then the downsampling process is performed layer by layer to obtain the label feature vector [z2, z3..., zk] corresponding to each label dimension.
在步骤S602中,预设函数为sigmoid函数,sigmoid函数可以表示为公式(4)所示:In step S602, the preset function is a sigmoid function, and the sigmoid function can be expressed as shown in formula (4):
Figure PCTCN2022090726-appb-000004
Figure PCTCN2022090726-appb-000004
通过sigmoid函数对标签特征向量进行识别,sigmoid函数会根据预设的画像标签类别对标签特征向量进行标签分类处理,在每一画像标签类别上创建一个概率分布,从而得到每一预设画像标签的概率值。Identify the label feature vector through the sigmoid function. The sigmoid function will classify the label feature vector according to the preset portrait label category, and create a probability distribution on each portrait label category, so as to obtain the value of each preset portrait label. probability value.
请参阅图7,在一些实施例,步骤S108还可以包括但不限于包括步骤S701至步骤S702:Please refer to FIG. 7, in some embodiments, step S108 may also include but not limited to include steps S701 to S702:
步骤S701,将概率值大于或者等于预设概率阈值的画像标签纳入同一集合,得到候选画像标签集;Step S701, including the portrait labels whose probability value is greater than or equal to the preset probability threshold into the same set to obtain a set of candidate portrait labels;
步骤S702,对候选画像标签集进行筛选处理,得到目标画像标签。Step S702, screening the candidate portrait label set to obtain the target portrait label.
具体地,首先执行步骤S701,若概率值小于预设概率阈值,则将概率值对应的画像标签进行过滤掉;若概率值大于或等于预设概率阈值,则将概率值对应的画像标签纳入候选画像标签集。例如,预设的概率阈值为0.6,当概率值大于或等于0.6时,则可以认为该用户拥有当前的画像标签。Specifically, first execute step S701, if the probability value is less than the preset probability threshold, filter out the portrait label corresponding to the probability value; if the probability value is greater than or equal to the preset probability threshold, then include the portrait label corresponding to the probability value into the candidate Portrait labels set. For example, the preset probability threshold is 0.6, and when the probability value is greater than or equal to 0.6, it can be considered that the user has the current portrait tag.
进一步地,执行步骤S702,可以采用人工复核等方式对候选画像标签集中的画像标签进行筛选,将与当前用户匹配度最高的画像标签提取出来,从而得到目标画像标签。另外,也 可以根据概率值的大小,对候选画像标签集内的画像标签进行降序排列,选取处于前五位的画像标签作为当前用户的目标画像标签。还可以是采用其他方式对候选画像标签集内的画像标签进行筛选,不限于此。Further, by executing step S702, the portrait tags in the candidate portrait tag set can be screened by means of manual review, etc., and the portrait tag with the highest matching degree with the current user is extracted to obtain the target portrait tag. In addition, the portrait tags in the candidate portrait tag set can also be arranged in descending order according to the size of the probability value, and the top five portrait tags are selected as the target portrait tags of the current user. Other methods may also be used to filter the portrait tags in the candidate portrait tag set, which is not limited thereto.
本申请实施例通过获取原始数据,其中,原始数据包括用户基础数据、用户行为数据以及用户评论数据。进而,对用户基础数据进行归一化处理,得到用户基础特征;通过预先训练的图卷积模型对用户行为数据进行特征提取,得到行为特征矩阵;对用户评论数据进行分词处理,得到评论文本词段向量,并将评论文本词段向量输入至预先训练的对比学习模型中,以使评论文本词段向量与对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量。这样一来,能够对不同类型的数据分别进行预处理,得到用户基础特征、行为特征矩阵以及评论词嵌入向量,提高了用户数据的合理性。进而,通过对用户基础特征、行为特征矩阵以及评论词嵌入向量进行融合处理,得到标准画像特征向量。最后,通过预先训练的标签识别模型对标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值,并根据概率值与预设概率阈值的大小关系,得到目标画像标签。本申请的多标签识别方法通过一个标签识别模型能够实现对不同画像标签的识别,相较于传统技术中需要针对不同标签类别分别训练分类器,能够极大地缩短模型训练时间,提高了识别效率。同时,本申请的多标签识别方法对于不同类型的用户数据分别进行了相应的数据预处理,使得获取到的标准画像特征向量能够更加符合识别需求,能够提高用户画像标签的识别准确性。In this embodiment of the present application, raw data is acquired, wherein the raw data includes basic user data, user behavior data, and user comment data. Furthermore, the user basic data is normalized to obtain the user basic features; the user behavior data is extracted through the pre-trained graph convolution model to obtain the behavior feature matrix; the user comment data is word-segmented to obtain the comment text word segment vector, and input the comment text segment vector into the pre-trained comparative learning model, so that the review text segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied to obtain the review word embedding vector. In this way, different types of data can be preprocessed separately to obtain user basic features, behavior feature matrices, and comment embedding vectors, which improves the rationality of user data. Furthermore, the standard portrait feature vector is obtained by fusing the user's basic features, behavior feature matrix, and comment word embedding vector. Finally, the pre-trained label recognition model is used to perform label recognition processing on the standard portrait feature vector to obtain the probability value of each preset portrait label, and obtain the target portrait label according to the relationship between the probability value and the preset probability threshold. The multi-label recognition method of the present application can realize the recognition of different portrait labels through a label recognition model. Compared with the traditional technology that needs to train classifiers for different label categories separately, it can greatly shorten the model training time and improve the recognition efficiency. At the same time, the multi-label recognition method of the present application performs corresponding data preprocessing for different types of user data, so that the obtained standard portrait feature vectors can better meet the recognition requirements, and can improve the recognition accuracy of user portrait labels.
请参阅图8,本申请实施例还提供一种多标签识别装置,可以实现上述多标签识别方法,该装置包括:Please refer to Figure 8, the embodiment of the present application also provides a multi-label identification device, which can realize the above-mentioned multi-label identification method, the device includes:
数据获取模块801,用于获取原始数据,其中,原始数据包括用户基础数据、用户行为数据以及用户评论数据;A data acquisition module 801, configured to acquire raw data, wherein the raw data includes user basic data, user behavior data and user comment data;
归一化模块802,用于对用户基础数据进行归一化处理,得到用户基础特征;A normalization module 802, configured to perform normalization processing on user basic data to obtain user basic features;
特征提取模块803,用于通过预先训练的图卷积模型对用户行为数据进行特征提取,得到行为特征矩阵;The feature extraction module 803 is used to perform feature extraction on user behavior data through a pre-trained graph convolution model to obtain a behavior feature matrix;
分词模块804,用于对用户评论数据进行分词处理,得到评论文本词段向量;The word segmentation module 804 is used to perform word segmentation processing on the user comment data to obtain the comment text word segment vector;
对比学习模块805,用于将评论文本词段向量输入至预先训练的对比学习模型中,以使评论文本词段向量与对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量; Contrastive learning module 805, for inputting the comment text word segment vector into the pre-trained comparative learning model, so that the comment text word segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied to obtain the comment word embedding vector ;
融合模块806,用于对用户基础特征、行为特征矩阵以及评论词嵌入向量进行融合处理,得到标准画像特征向量;The fusion module 806 is used to fuse the user basic features, behavior feature matrix and comment word embedding vectors to obtain standard portrait feature vectors;
标签识别模块807,用于通过预先训练的标签识别模型对标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值;The label recognition module 807 is used to perform label recognition processing on the standard portrait feature vector through the pre-trained label recognition model to obtain the probability value of each preset portrait label;
比较模块808,用于根据概率值与预设概率阈值的大小关系,得到目标画像标签。The comparison module 808 is configured to obtain the target portrait label according to the magnitude relationship between the probability value and the preset probability threshold.
该多标签识别装置的具体实施方式与上述多标签识别方法的具体实施例基本相同,在此不再赘述。The specific implementation of the multi-tag identification device is basically the same as the specific embodiment of the above-mentioned multi-label identification method, and will not be repeated here.
本申请实施例还提供了一种电子设备,电子设备包括:存储器、处理器、存储在存储器上并可在处理器上运行的程序以及用于实现处理器和存储器之间的连接通信的数据总线,程序被处理器执行时实现上述多标签识别方法。该电子设备可以为包括平板电脑、车载电脑等任意智能终端。The embodiment of the present application also provides an electronic device, the electronic device includes: a memory, a processor, a program stored in the memory and operable on the processor, and a data bus for realizing connection and communication between the processor and the memory , when the program is executed by the processor, the above multi-label identification method is realized. The electronic device may be any intelligent terminal including a tablet computer, a vehicle-mounted computer, and the like.
请参阅图9,图9示意了另一实施例的电子设备的硬件结构,电子设备包括:Please refer to FIG. 9. FIG. 9 illustrates a hardware structure of an electronic device in another embodiment. The electronic device includes:
处理器901,可以采用通用的CPU(CentralProcessingUnit,中央处理器)、微处理器、应用专用集成电路(ApplicationSpecificIntegratedCircuit,ASIC)、或者一个或多个集成电路等方式实现,用于执行相关程序,以实现本申请实施例所提供的技术方案;The processor 901 may be implemented by a general-purpose CPU (Central Processing Unit, central processing unit), a microprocessor, an application-specific integrated circuit (Application Specific Integrated Circuit, ASIC), or one or more integrated circuits, and is used to execute related programs, so as to realize The technical solutions provided by the embodiments of the present application;
存储器902,可以采用只读存储器(ReadOnlyMemory,ROM)、静态存储设备、动态存储设备或者随机存取存储器(RandomAccessMemory,RAM)等形式实现。存储器902可以存储操作系统和其他应用程序,在通过软件或者固件来实现本说明书实施例所提供的技术方案时,相关的程序代码保存在存储器902中,并由处理器901来调用执行一种多标签识别方法,其中, 多标签识别方法包括:获取原始数据,其中,原始数据包括用户基础数据、用户行为数据以及用户评论数据;对用户基础数据进行归一化处理,得到用户基础特征;通过预先训练的图卷积模型对用户行为数据进行特征提取,得到行为特征矩阵;对用户评论数据进行分词处理,得到评论文本词段向量;将评论文本词段向量输入至预先训练的对比学习模型中,以使评论文本词段向量与对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量;对用户基础特征、行为特征矩阵以及评论词嵌入向量进行融合处理,得到标准画像特征向量;通过预先训练的标签识别模型对标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值;根据概率值与预设概率阈值的大小关系,得到目标画像标签;The memory 902 may be implemented in the form of a read-only memory (ReadOnlyMemory, ROM), a static storage device, a dynamic storage device, or a random access memory (RandomAccessMemory, RAM). The memory 902 can store operating systems and other application programs. When implementing the technical solutions provided by the embodiments of this specification through software or firmware, the relevant program codes are stored in the memory 902 and called by the processor 901 to execute a multi- The label recognition method, wherein the multi-label recognition method includes: obtaining original data, wherein the original data includes user basic data, user behavior data, and user comment data; normalizing the user basic data to obtain user basic features; The trained graph convolution model extracts features from user behavior data to obtain a behavior feature matrix; performs word segmentation processing on user comment data to obtain comment text segment vectors; input comment text segment vectors into the pre-trained comparative learning model, Multiply the word segment vector of the comment text with the reference word embedding matrix in the comparative learning model to obtain the comment word embedding vector; perform fusion processing on the user basic features, behavior feature matrix and comment word embedding vector to obtain the standard portrait feature vector Carrying out label recognition processing on the standard portrait feature vector through the pre-trained label recognition model to obtain the probability value of each preset portrait label; according to the size relationship between the probability value and the preset probability threshold, the target portrait label is obtained;
输入/输出接口903,用于实现信息输入及输出;The input/output interface 903 is used to realize information input and output;
通信接口904,用于实现本设备与其他设备的通信交互,可以通过有线方式(例如USB、网线等)实现通信,也可以通过无线方式(例如移动网络、WIFI、蓝牙等)实现通信;The communication interface 904 is used to realize the communication interaction between the device and other devices, and the communication can be realized through a wired method (such as USB, network cable, etc.), or can be realized through a wireless method (such as a mobile network, WIFI, Bluetooth, etc.);
总线905,在设备的各个组件(例如处理器901、存储器902、输入/输出接口903和通信接口904)之间传输信息;bus 905, for transferring information between various components of the device (such as processor 901, memory 902, input/output interface 903 and communication interface 904);
其中处理器901、存储器902、输入/输出接口903和通信接口904通过总线905实现彼此之间在设备内部的通信连接。The processor 901 , the memory 902 , the input/output interface 903 and the communication interface 904 are connected to each other within the device through the bus 905 .
本申请实施例还提供了一种存储介质,存储介质为计算机可读存储介质,用于计算机可读存储,所述计算机可读存储介质可以是非易失性,也可以是易失性。存储介质存储有一个或者多个程序,一个或者多个程序可被一个或者多个处理器执行,以实现一种多标签识别方法,其中,多标签识别方法包括:获取原始数据,其中,原始数据包括用户基础数据、用户行为数据以及用户评论数据;对用户基础数据进行归一化处理,得到用户基础特征;通过预先训练的图卷积模型对用户行为数据进行特征提取,得到行为特征矩阵;对用户评论数据进行分词处理,得到评论文本词段向量;将评论文本词段向量输入至预先训练的对比学习模型中,以使评论文本词段向量与对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量;对用户基础特征、行为特征矩阵以及评论词嵌入向量进行融合处理,得到标准画像特征向量;通过预先训练的标签识别模型对标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值;根据概率值与预设概率阈值的大小关系,得到目标画像标签。An embodiment of the present application also provides a storage medium, which is a computer-readable storage medium for computer-readable storage, and the computer-readable storage medium may be non-volatile or volatile. The storage medium stores one or more programs, and the one or more programs can be executed by one or more processors to implement a multi-label identification method, wherein the multi-label identification method includes: obtaining original data, wherein the original data Including user basic data, user behavior data and user comment data; normalize user basic data to obtain user basic features; use pre-trained graph convolution model to extract user behavior data to obtain behavior feature matrix; The user comment data is subjected to word segmentation processing to obtain the comment text segment vector; the comment text segment vector is input into the pre-trained comparative learning model, so that the comment text segment vector and the reference word embedding matrix in the comparative learning model are matrix correlated. Multiply the embedding vector of the comment to get the embedding vector of the comment; perform fusion processing on the user’s basic features, behavior feature matrix and embedding vector of the comment to obtain the standard portrait feature vector; use the pre-trained label recognition model to perform label recognition on the standard portrait feature vector to obtain each A probability value of a preset portrait label; according to a magnitude relationship between the probability value and a preset probability threshold, a target portrait label is obtained.
存储器作为一种非暂态计算机可读存储介质,可用于存储非暂态软件程序以及非暂态性计算机可执行程序。此外,存储器可以包括高速随机存取存储器,还可以包括非暂态存储器,例如至少一个磁盘存储器件、闪存器件、或其他非暂态固态存储器件。在一些实施方式中,存储器可选包括相对于处理器远程设置的存储器,这些远程存储器可以通过网络连接至该处理器。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。As a non-transitory computer-readable storage medium, memory can be used to store non-transitory software programs and non-transitory computer-executable programs. In addition, the memory may include high-speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid-state storage devices. In some embodiments, the memory optionally includes memory located remotely from the processor, and these remote memories may be connected to the processor via a network. Examples of the aforementioned networks include, but are not limited to, the Internet, intranets, local area networks, mobile communication networks, and combinations thereof.
本申请实施例描述的实施例是为了更加清楚的说明本申请实施例的技术方案,并不构成对于本申请实施例提供的技术方案的限定,本领域技术人员可知,随着技术的演变和新应用场景的出现,本申请实施例提供的技术方案对于类似的技术问题,同样适用。The embodiments described in the embodiments of the present application are to illustrate the technical solutions of the embodiments of the present application more clearly, and do not constitute a limitation to the technical solutions provided by the embodiments of the present application. Those skilled in the art know that with the evolution of technology and new For the emergence of application scenarios, the technical solutions provided by the embodiments of the present application are also applicable to similar technical problems.
本领域技术人员可以理解的是,图1-7中示出的技术方案并不构成对本申请实施例的限定,可以包括比图示更多或更少的步骤,或者组合某些步骤,或者不同的步骤。Those skilled in the art can understand that the technical solutions shown in Figures 1-7 do not constitute a limitation to the embodiments of the present application, and may include more or fewer steps than those shown in the illustrations, or combine certain steps, or be different A step of.
以上所描述的装置实施例仅仅是示意性的,其中作为分离部件说明的单元可以是或者也可以不是物理上分开的,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部模块来实现本实施例方案的目的。The device embodiments described above are only illustrative, and the units described as separate components may or may not be physically separated, that is, they may be located in one place, or may be distributed to multiple network units. Part or all of the modules can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
本领域普通技术人员可以理解,上文中所公开方法中的全部或某些步骤、系统、设备中的功能模块/单元可以被实施为软件、固件、硬件及其适当的组合。Those of ordinary skill in the art can understand that all or some of the steps in the methods disclosed above, the functional modules/units in the system, and the device can be implemented as software, firmware, hardware, and an appropriate combination thereof.
另外,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, each functional unit in each embodiment of the present application may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit. The above-mentioned integrated units can be implemented in the form of hardware or in the form of software functional units.
集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请的技术方案本质上或者说对现有 技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括多指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本申请各个实施例的方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(Read-Only Memory,简称ROM)、随机存取存储器(Random Access Memory,简称RAM)、磁碟或者光盘等各种可以存储程序的介质。If the integrated unit is realized in the form of a software function unit and sold or used as an independent product, it can be stored in a computer-readable storage medium. Based on this understanding, the technical solution of the present application is essentially or part of the contribution to the prior art or all or part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium , including multiple instructions to make a computer device (which may be a personal computer, a server, or a network device, etc.) execute all or part of the steps of the method in each embodiment of the present application. The aforementioned storage media include: U disk, mobile hard disk, read-only memory (Read-Only Memory, referred to as ROM), random access memory (Random Access Memory, referred to as RAM), magnetic disk or optical disc, etc., which can store programs. medium.
以上参照附图说明了本申请实施例的优选实施例,并非因此局限本申请实施例的权利范围。本领域技术人员不脱离本申请实施例的范围和实质内所作的任何修改、等同替换和改进,均应在本申请实施例的权利范围之内。The preferred embodiments of the embodiments of the present application have been described above with reference to the accompanying drawings, which does not limit the scope of rights of the embodiments of the present application. Any modifications, equivalent replacements and improvements made by those skilled in the art without departing from the scope and essence of the embodiments of the present application shall fall within the scope of rights of the embodiments of the present application.

Claims (20)

  1. 一种多标签识别方法,其中,所述方法包括:A multi-label identification method, wherein the method comprises:
    获取原始数据,其中,所述原始数据包括用户基础数据、用户行为数据以及用户评论数据;Obtaining raw data, wherein the raw data includes user basic data, user behavior data and user comment data;
    对所述用户基础数据进行归一化处理,得到用户基础特征;performing normalization processing on the user basic data to obtain user basic features;
    通过预先训练的图卷积模型对所述用户行为数据进行特征提取,得到行为特征矩阵;performing feature extraction on the user behavior data through a pre-trained graph convolution model to obtain a behavior feature matrix;
    对所述用户评论数据进行分词处理,得到评论文本词段向量;Carry out word segmentation processing to described user comment data, obtain comment text word segment vector;
    将所述评论文本词段向量输入至预先训练的对比学习模型中,以使所述评论文本词段向量与所述对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量;The comment text word segment vector is input into the pre-trained comparative learning model, so that the comment text word segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied to obtain the comment word embedding vector;
    对所述用户基础特征、所述行为特征矩阵以及所述评论词嵌入向量进行融合处理,得到标准画像特征向量;Perform fusion processing on the user basic features, the behavior feature matrix and the comment word embedding vector to obtain a standard portrait feature vector;
    通过预先训练的标签识别模型对所述标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值;Carrying out label recognition processing on the standard portrait feature vector through a pre-trained label recognition model to obtain the probability value of each preset portrait label;
    根据所述概率值与预设概率阈值的大小关系,得到目标画像标签。According to the magnitude relationship between the probability value and the preset probability threshold, the target portrait label is obtained.
  2. 根据权利要求1所述的多标签识别方法,其中,所述通过预先训练的图卷积模型对所述用户行为数据进行特征提取,得到行为特征矩阵的步骤,包括:The multi-label recognition method according to claim 1, wherein the step of performing feature extraction on the user behavior data through a pre-trained graph convolution model to obtain a behavior feature matrix includes:
    将所述用户行为数据映射到预设的向量空间,得到用户行为特征向量;Mapping the user behavior data to a preset vector space to obtain a user behavior feature vector;
    根据预设的课程类型和所述用户行为特征向量,构建行为特征图;Construct a behavior feature map according to the preset course type and the user behavior feature vector;
    对所述行为特征图进行图卷积处理,得到行为度矩阵和行为邻接矩阵;Carrying out graph convolution processing on the behavior feature map to obtain a behavior degree matrix and a behavior adjacency matrix;
    对所述行为度矩阵和所述行为矩阵进行做差处理,得到行为特征矩阵。Performing difference processing on the behavior degree matrix and the behavior matrix to obtain a behavior feature matrix.
  3. 根据权利要求1所述的多标签识别方法,其中,所述对所述用户评论数据进行分词处理,得到评论文本词段向量的步骤,包括:The multi-label recognition method according to claim 1, wherein the step of performing word segmentation processing on the user comment data to obtain a comment text word segment vector includes:
    通过预设的分词器对所述用户评论数据进行分词处理,得到评论文本词段;Segmenting the user comment data through a preset word segmenter to obtain comment text segments;
    对所述评论文本词段进行编码处理,得到评论文本词段向量。Encoding processing is performed on the word segment of the comment text to obtain a word segment vector of the comment text.
  4. 根据权利要求1所述的多标签识别方法,其中,所述将所述评论文本词段向量输入至预先训练的对比学习模型中,以使所述评论文本词段向量与所述对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量的步骤,包括:The multi-label recognition method according to claim 1, wherein, the comment text word segment vector is input into a pre-trained contrastive learning model, so that the comment text word segment vector is the same as that in the contrastive learning model The reference word embedding matrix is multiplied by matrix, and the steps of obtaining the comment word embedding vector include:
    将所述评论文本词段向量输入到对比学习模型中,以使所述评论文本词段向量与参考词嵌入矩阵进行矩阵相乘,得到多个基本词嵌入向量;The comment text word segment vector is input into the comparative learning model, so that the comment text word segment vector and the reference word embedding matrix are matrix multiplied to obtain a plurality of basic word embedding vectors;
    对所述基本词嵌入向量进行映射处理,得到评论词嵌入向量。The basic word embedding vector is mapped to obtain the comment word embedding vector.
  5. 根据权利要求1所述的多标签识别方法,其中,所述通过预先训练的标签识别模型对所述标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值的步骤,包括:The multi-label recognition method according to claim 1, wherein the step of performing label recognition processing on the standard portrait feature vector through a pre-trained label recognition model to obtain the probability value of each preset portrait label includes:
    根据预设的标签维度对所述标准画像特征向量进行重构处理,得到标签特征向量;Reconstructing the standard portrait feature vector according to the preset label dimension to obtain the label feature vector;
    利用预设函数对所述标签特征向量进行识别处理,得到每一预设画像标签的概率值。The label feature vector is identified by using a preset function to obtain the probability value of each preset portrait label.
  6. 根据权利要求1所述的多标签识别方法,其中,所述根据所述概率值与预设概率阈值的大小关系,得到目标画像标签的步骤,包括:The multi-label identification method according to claim 1, wherein the step of obtaining the target portrait label according to the magnitude relationship between the probability value and the preset probability threshold comprises:
    将所述概率值大于或者等于所述预设概率阈值的画像标签纳入同一集合,得到候选画像标签集;Incorporating portrait tags whose probability values are greater than or equal to the preset probability threshold into the same set to obtain a set of candidate portrait tags;
    对所述候选画像标签集进行筛选处理,得到所述目标画像标签。The candidate portrait label set is screened to obtain the target portrait label.
  7. 根据权利要求1至6任一项所述的多标签识别方法,其中,在所述通过预先训练的标签识别模型对所述标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值的步骤之前,所述方法还包括预先训练所述标签识别模型,具体包括:The multi-label recognition method according to any one of claims 1 to 6, wherein, performing label recognition processing on the standard portrait feature vector through the pre-trained label recognition model to obtain the probability of each preset portrait label Before the step of value, the method also includes pre-training the label recognition model, specifically including:
    获取标注用户数据;Obtain marked user data;
    对所述标注用户数据进行特征提取,得到样本特征向量;performing feature extraction on the labeled user data to obtain a sample feature vector;
    将所述样本特征向量输入到标签识别模型中;The sample feature vector is input into the label recognition model;
    通过所述标签识别模型的损失函数计算出每一画像标签类别的样本概率预测值;Calculate the sample probability prediction value of each portrait label category through the loss function of the label recognition model;
    根据所述样本概率预测值对所述标签识别模型的损失函数进行优化,以更新所述标签识别模型。Optimizing the loss function of the label recognition model according to the sample probability prediction value, so as to update the label recognition model.
  8. 一种多标签识别装置,其中,所述装置包括:A multi-label identification device, wherein the device includes:
    数据获取模块,用于获取原始数据,其中,所述原始数据包括用户基础数据、用户行为数据以及用户评论数据;A data acquisition module, configured to acquire raw data, wherein the raw data includes user basic data, user behavior data and user comment data;
    归一化模块,用于对所述用户基础数据进行归一化处理,得到用户基础特征;A normalization module, configured to perform normalization processing on the user basic data to obtain user basic features;
    特征提取模块,用于通过预先训练的图卷积模型对所述用户行为数据进行特征提取,得到行为特征矩阵;Feature extraction module, for carrying out feature extraction to described user behavior data by pre-trained graph convolution model, obtains behavior feature matrix;
    分词模块,用于对所述用户评论数据进行分词处理,得到评论文本词段向量;A word segmentation module, configured to perform word segmentation processing on the user comment data to obtain a comment text word segment vector;
    对比学习模块,用于将所述评论文本词段向量输入至预先训练的对比学习模型中,以使所述评论文本词段向量与所述对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量;A comparative learning module, for inputting the comment text word segment vector into a pre-trained comparative learning model, so that the comment text word segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied, Get the comment embedding vector;
    融合模块,用于对所述用户基础特征、所述行为特征矩阵以及所述评论词嵌入向量进行融合处理,得到标准画像特征向量;The fusion module is used to perform fusion processing on the user basic features, the behavior feature matrix and the comment word embedding vector to obtain a standard portrait feature vector;
    标签识别模块,用于通过预先训练的标签识别模型对所述标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值;A tag recognition module, configured to perform tag recognition processing on the standard portrait feature vector through a pre-trained tag recognition model to obtain the probability value of each preset portrait tag;
    比较模块,用于根据所述概率值与预设概率阈值的大小关系,得到目标画像标签。The comparison module is used to obtain the target portrait label according to the magnitude relationship between the probability value and the preset probability threshold.
  9. 一种电子设备,其中,所述电子设备包括存储器、处理器、存储在所述存储器上并可在所述处理器上运行的程序以及用于实现所述处理器和所述存储器之间的连接通信的数据总线,所述程序被所述处理器执行时实现如下步骤:An electronic device, wherein the electronic device includes a memory, a processor, a program stored on the memory and operable on the processor, and a program for realizing the connection between the processor and the memory A data bus for communication, when the program is executed by the processor, the following steps are implemented:
    获取原始数据,其中,所述原始数据包括用户基础数据、用户行为数据以及用户评论数据;Obtaining raw data, wherein the raw data includes user basic data, user behavior data and user comment data;
    对所述用户基础数据进行归一化处理,得到用户基础特征;performing normalization processing on the user basic data to obtain user basic features;
    通过预先训练的图卷积模型对所述用户行为数据进行特征提取,得到行为特征矩阵;performing feature extraction on the user behavior data through a pre-trained graph convolution model to obtain a behavior feature matrix;
    对所述用户评论数据进行分词处理,得到评论文本词段向量;Carry out word segmentation processing to described user comment data, obtain comment text word segment vector;
    将所述评论文本词段向量输入至预先训练的对比学习模型中,以使所述评论文本词段向量与所述对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量;The comment text word segment vector is input into the pre-trained comparative learning model, so that the comment text word segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied to obtain the comment word embedding vector;
    对所述用户基础特征、所述行为特征矩阵以及所述评论词嵌入向量进行融合处理,得到标准画像特征向量;Perform fusion processing on the user basic features, the behavior feature matrix and the comment word embedding vector to obtain a standard portrait feature vector;
    通过预先训练的标签识别模型对所述标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值;Carrying out label recognition processing on the standard portrait feature vector through a pre-trained label recognition model to obtain the probability value of each preset portrait label;
    根据所述概率值与预设概率阈值的大小关系,得到目标画像标签。According to the magnitude relationship between the probability value and the preset probability threshold, the target portrait label is obtained.
  10. 根据权利要求9所述的电子设备,其中,所述通过预先训练的图卷积模型对所述用户行为数据进行特征提取,得到行为特征矩阵的步骤,包括:The electronic device according to claim 9, wherein the step of performing feature extraction on the user behavior data through a pre-trained graph convolution model to obtain a behavior feature matrix includes:
    将所述用户行为数据映射到预设的向量空间,得到用户行为特征向量;Mapping the user behavior data to a preset vector space to obtain a user behavior feature vector;
    根据预设的课程类型和所述用户行为特征向量,构建行为特征图;Construct a behavior feature map according to the preset course type and the user behavior feature vector;
    对所述行为特征图进行图卷积处理,得到行为度矩阵和行为邻接矩阵;Carrying out graph convolution processing on the behavior feature map to obtain a behavior degree matrix and a behavior adjacency matrix;
    对所述行为度矩阵和所述行为矩阵进行做差处理,得到行为特征矩阵。Performing difference processing on the behavior degree matrix and the behavior matrix to obtain a behavior feature matrix.
  11. 根据权利要求9所述的电子设备,其中,所述对所述用户评论数据进行分词处理,得到评论文本词段向量的步骤,包括:The electronic device according to claim 9, wherein the step of performing word segmentation processing on the user comment data to obtain the comment text word segment vector comprises:
    通过预设的分词器对所述用户评论数据进行分词处理,得到评论文本词段;Segmenting the user comment data through a preset word segmenter to obtain comment text segments;
    对所述评论文本词段进行编码处理,得到评论文本词段向量。Encoding processing is performed on the word segment of the comment text to obtain a word segment vector of the comment text.
  12. 根据权利要求9所述的电子设备,其中,所述将所述评论文本词段向量输入至预先训练的对比学习模型中,以使所述评论文本词段向量与所述对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量的步骤,包括:The electronic device according to claim 9 , wherein the comment text word segment vector is input into a pre-trained comparative learning model, so that the comment text word segment vector and the reference in the comparative learning model The word embedding matrix is multiplied by matrix to obtain the steps of comment word embedding vector, including:
    将所述评论文本词段向量输入到对比学习模型中,以使所述评论文本词段向量与参考词嵌入矩阵进行矩阵相乘,得到多个基本词嵌入向量;The comment text word segment vector is input into the comparative learning model, so that the comment text word segment vector and the reference word embedding matrix are matrix multiplied to obtain a plurality of basic word embedding vectors;
    对所述基本词嵌入向量进行映射处理,得到评论词嵌入向量。The basic word embedding vector is mapped to obtain the comment word embedding vector.
  13. 根据权利要求9所述的电子设备,其中,所述通过预先训练的标签识别模型对所述标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值的步骤,包括:The electronic device according to claim 9, wherein the step of performing label recognition processing on the standard portrait feature vector through a pre-trained label recognition model to obtain the probability value of each preset portrait label comprises:
    根据预设的标签维度对所述标准画像特征向量进行重构处理,得到标签特征向量;Reconstructing the standard portrait feature vector according to the preset label dimension to obtain the label feature vector;
    利用预设函数对所述标签特征向量进行识别处理,得到每一预设画像标签的概率值。The label feature vector is identified by using a preset function to obtain the probability value of each preset portrait label.
  14. 根据权利要求9所述的电子设备,其中,所述根据所述概率值与预设概率阈值的大小关系,得到目标画像标签的步骤,包括:The electronic device according to claim 9, wherein the step of obtaining the target portrait label according to the magnitude relationship between the probability value and the preset probability threshold comprises:
    将所述概率值大于或者等于所述预设概率阈值的画像标签纳入同一集合,得到候选画像标签集;Incorporating portrait tags whose probability values are greater than or equal to the preset probability threshold into the same set to obtain a set of candidate portrait tags;
    对所述候选画像标签集进行筛选处理,得到所述目标画像标签。The candidate portrait label set is screened to obtain the target portrait label.
  15. 一种存储介质,所述存储介质为计算机可读存储介质,用于计算机可读存储,其中,所述存储介质存储有一个或者多个程序,所述一个或者多个程序可被一个或者多个处理器执行,以实现如下步骤:A storage medium, the storage medium is a computer-readable storage medium for computer-readable storage, wherein the storage medium stores one or more programs, and the one or more programs can be used by one or more The processor executes to realize the following steps:
    获取原始数据,其中,所述原始数据包括用户基础数据、用户行为数据以及用户评论数据;Obtaining raw data, wherein the raw data includes user basic data, user behavior data and user comment data;
    对所述用户基础数据进行归一化处理,得到用户基础特征;performing normalization processing on the user basic data to obtain user basic features;
    通过预先训练的图卷积模型对所述用户行为数据进行特征提取,得到行为特征矩阵;performing feature extraction on the user behavior data through a pre-trained graph convolution model to obtain a behavior feature matrix;
    对所述用户评论数据进行分词处理,得到评论文本词段向量;Carry out word segmentation processing to described user comment data, obtain comment text word segment vector;
    将所述评论文本词段向量输入至预先训练的对比学习模型中,以使所述评论文本词段向量与所述对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量;The comment text word segment vector is input into the pre-trained comparative learning model, so that the comment text word segment vector and the reference word embedding matrix in the comparative learning model are matrix multiplied to obtain the comment word embedding vector;
    对所述用户基础特征、所述行为特征矩阵以及所述评论词嵌入向量进行融合处理,得到标准画像特征向量;Perform fusion processing on the user basic features, the behavior feature matrix and the comment word embedding vector to obtain a standard portrait feature vector;
    通过预先训练的标签识别模型对所述标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值;Carrying out label recognition processing on the standard portrait feature vector through a pre-trained label recognition model to obtain the probability value of each preset portrait label;
    根据所述概率值与预设概率阈值的大小关系,得到目标画像标签。According to the magnitude relationship between the probability value and the preset probability threshold, the target portrait label is obtained.
  16. 根据权利要求15所述的存储介质,其中,所述通过预先训练的图卷积模型对所述用户行为数据进行特征提取,得到行为特征矩阵的步骤,包括:The storage medium according to claim 15, wherein the step of performing feature extraction on the user behavior data through a pre-trained graph convolution model to obtain a behavior feature matrix includes:
    将所述用户行为数据映射到预设的向量空间,得到用户行为特征向量;Mapping the user behavior data to a preset vector space to obtain a user behavior feature vector;
    根据预设的课程类型和所述用户行为特征向量,构建行为特征图;Construct a behavior feature map according to the preset course type and the user behavior feature vector;
    对所述行为特征图进行图卷积处理,得到行为度矩阵和行为邻接矩阵;Carrying out graph convolution processing on the behavior feature map to obtain a behavior degree matrix and a behavior adjacency matrix;
    对所述行为度矩阵和所述行为矩阵进行做差处理,得到行为特征矩阵。Performing difference processing on the behavior degree matrix and the behavior matrix to obtain a behavior feature matrix.
  17. 根据权利要求15所述的存储介质,其中,所述对所述用户评论数据进行分词处理,得到评论文本词段向量的步骤,包括:The storage medium according to claim 15, wherein the step of performing word segmentation processing on the user comment data to obtain the comment text word segment vector comprises:
    通过预设的分词器对所述用户评论数据进行分词处理,得到评论文本词段;Segmenting the user comment data through a preset word segmenter to obtain comment text segments;
    对所述评论文本词段进行编码处理,得到评论文本词段向量。Encoding processing is performed on the word segment of the comment text to obtain a word segment vector of the comment text.
  18. 根据权利要求15所述的存储介质,其中,所述将所述评论文本词段向量输入至预先训练的对比学习模型中,以使所述评论文本词段向量与所述对比学习模型中的参考词嵌入矩阵进行矩阵相乘,得到评论词嵌入向量的步骤,包括:The storage medium according to claim 15, wherein the comment text word segment vector is input into a pre-trained comparative learning model, so that the comment text word segment vector and the reference in the comparative learning model The word embedding matrix is multiplied by matrix to obtain the steps of comment word embedding vector, including:
    将所述评论文本词段向量输入到对比学习模型中,以使所述评论文本词段向量与参考词嵌入矩阵进行矩阵相乘,得到多个基本词嵌入向量;The comment text word segment vector is input into the comparative learning model, so that the comment text word segment vector and the reference word embedding matrix are matrix multiplied to obtain a plurality of basic word embedding vectors;
    对所述基本词嵌入向量进行映射处理,得到评论词嵌入向量。The basic word embedding vector is mapped to obtain the comment word embedding vector.
  19. 根据权利要求15所述的存储介质,其中,所述通过预先训练的标签识别模型对所述标准画像特征向量进行标签识别处理,得到每一预设画像标签的概率值的步骤,包括:The storage medium according to claim 15, wherein the step of performing label recognition processing on the standard portrait feature vector through a pre-trained label recognition model to obtain the probability value of each preset portrait label includes:
    根据预设的标签维度对所述标准画像特征向量进行重构处理,得到标签特征向量;Reconstructing the standard portrait feature vector according to the preset label dimension to obtain the label feature vector;
    利用预设函数对所述标签特征向量进行识别处理,得到每一预设画像标签的概率值。The label feature vector is identified by using a preset function to obtain the probability value of each preset portrait label.
  20. 根据权利要求15所述的存储介质,其中,所述根据所述概率值与预设概率阈值的大小关系,得到目标画像标签的步骤,包括:The storage medium according to claim 15, wherein the step of obtaining the target portrait label according to the magnitude relationship between the probability value and the preset probability threshold comprises:
    将所述概率值大于或者等于所述预设概率阈值的画像标签纳入同一集合,得到候选画像标签集;Incorporating portrait tags whose probability values are greater than or equal to the preset probability threshold into the same set to obtain a set of candidate portrait tags;
    对所述候选画像标签集进行筛选处理,得到所述目标画像标签。The candidate portrait label set is screened to obtain the target portrait label.
PCT/CN2022/090726 2022-01-11 2022-04-29 Multi-label identification method and apparatus, electronic device, and storage medium WO2023134084A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210027793.0 2022-01-11
CN202210027793.0A CN114358007A (en) 2022-01-11 2022-01-11 Multi-label identification method and device, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
WO2023134084A1 true WO2023134084A1 (en) 2023-07-20

Family

ID=81108800

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/090726 WO2023134084A1 (en) 2022-01-11 2022-04-29 Multi-label identification method and apparatus, electronic device, and storage medium

Country Status (2)

Country Link
CN (1) CN114358007A (en)
WO (1) WO2023134084A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116910628A (en) * 2023-09-12 2023-10-20 联通在线信息科技有限公司 Creator expertise portrait assessment method and system
CN116910377A (en) * 2023-09-14 2023-10-20 长威信息科技发展股份有限公司 Grid event classified search recommendation method and system

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114358007A (en) * 2022-01-11 2022-04-15 平安科技(深圳)有限公司 Multi-label identification method and device, electronic equipment and storage medium
CN115035763A (en) * 2022-06-22 2022-09-09 深圳市沃特沃德信息有限公司 Dictation optimization method and device, computer equipment and storage medium
CN115689648B (en) * 2022-10-28 2023-07-28 广东柏烨互动网络科技有限公司 User information processing method and system applied to directional delivery
CN117421497B (en) * 2023-11-02 2024-04-26 北京蜂鸟映像电子商务有限公司 Work object processing method and device, readable storage medium and electronic equipment

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200142999A1 (en) * 2018-11-02 2020-05-07 Valve Corporation Classification and moderation of text
CN112528163A (en) * 2020-12-04 2021-03-19 中山大学 Social platform user occupation prediction method based on graph convolution network
CN113139141A (en) * 2021-04-22 2021-07-20 康键信息技术(深圳)有限公司 User label extension labeling method, device, equipment and storage medium
CN113792818A (en) * 2021-10-18 2021-12-14 平安科技(深圳)有限公司 Intention classification method and device, electronic equipment and computer readable storage medium
CN113868417A (en) * 2021-09-27 2021-12-31 平安国际智慧城市科技股份有限公司 Sensitive comment identification method and device, terminal equipment and storage medium
CN113887215A (en) * 2021-10-18 2022-01-04 平安科技(深圳)有限公司 Text similarity calculation method and device, electronic equipment and storage medium
CN114358007A (en) * 2022-01-11 2022-04-15 平安科技(深圳)有限公司 Multi-label identification method and device, electronic equipment and storage medium

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109918507B (en) * 2019-03-08 2021-04-27 北京工业大学 textCNN (text-based network communication network) improved text classification method
CN113392317A (en) * 2021-01-07 2021-09-14 腾讯科技(深圳)有限公司 Label configuration method, device, equipment and storage medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200142999A1 (en) * 2018-11-02 2020-05-07 Valve Corporation Classification and moderation of text
CN112528163A (en) * 2020-12-04 2021-03-19 中山大学 Social platform user occupation prediction method based on graph convolution network
CN113139141A (en) * 2021-04-22 2021-07-20 康键信息技术(深圳)有限公司 User label extension labeling method, device, equipment and storage medium
CN113868417A (en) * 2021-09-27 2021-12-31 平安国际智慧城市科技股份有限公司 Sensitive comment identification method and device, terminal equipment and storage medium
CN113792818A (en) * 2021-10-18 2021-12-14 平安科技(深圳)有限公司 Intention classification method and device, electronic equipment and computer readable storage medium
CN113887215A (en) * 2021-10-18 2022-01-04 平安科技(深圳)有限公司 Text similarity calculation method and device, electronic equipment and storage medium
CN114358007A (en) * 2022-01-11 2022-04-15 平安科技(深圳)有限公司 Multi-label identification method and device, electronic equipment and storage medium

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116910628A (en) * 2023-09-12 2023-10-20 联通在线信息科技有限公司 Creator expertise portrait assessment method and system
CN116910628B (en) * 2023-09-12 2024-02-06 联通在线信息科技有限公司 Creator expertise portrait assessment method and system
CN116910377A (en) * 2023-09-14 2023-10-20 长威信息科技发展股份有限公司 Grid event classified search recommendation method and system
CN116910377B (en) * 2023-09-14 2023-12-08 长威信息科技发展股份有限公司 Grid event classified search recommendation method and system

Also Published As

Publication number Publication date
CN114358007A (en) 2022-04-15

Similar Documents

Publication Publication Date Title
WO2023134084A1 (en) Multi-label identification method and apparatus, electronic device, and storage medium
EP3660733B1 (en) Method and system for information extraction from document images using conversational interface and database querying
CN111582409B (en) Training method of image tag classification network, image tag classification method and device
CN108694225B (en) Image searching method, feature vector generating method and device and electronic equipment
US20210382937A1 (en) Image processing method and apparatus, and storage medium
CN105354307B (en) Image content identification method and device
CN110969020A (en) CNN and attention mechanism-based Chinese named entity identification method, system and medium
CN112380435A (en) Literature recommendation method and recommendation system based on heterogeneous graph neural network
CN112613308A (en) User intention identification method and device, terminal equipment and storage medium
CN111475622A (en) Text classification method, device, terminal and storage medium
WO2023134088A1 (en) Video summary generation method and apparatus, electronic device, and storage medium
CN111444715B (en) Entity relationship identification method and device, computer equipment and storage medium
CN111159485A (en) Tail entity linking method, device, server and storage medium
WO2023179429A1 (en) Video data processing method and apparatus, electronic device, and storage medium
CN112580328A (en) Event information extraction method and device, storage medium and electronic equipment
CN113486175B (en) Text classification method, text classification device, computer device, and storage medium
CN114612921B (en) Form recognition method and device, electronic equipment and computer readable medium
CN111191031A (en) Entity relation classification method of unstructured text based on WordNet and IDF
CN113836992A (en) Method for identifying label, method, device and equipment for training label identification model
CN108805280B (en) Image retrieval method and device
US20200167655A1 (en) Method and apparatus for re-configuring neural network
CN112364166B (en) Method for establishing relation extraction model and relation extraction method
CN113536784A (en) Text processing method and device, computer equipment and storage medium
CN116775875A (en) Question corpus construction method and device, question answering method and device and storage medium
CN115828153A (en) Task prediction method, device, equipment and medium based on artificial intelligence

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22919724

Country of ref document: EP

Kind code of ref document: A1