CN111930930A - Abstract comment abstract generation method based on commodity aspect alignment - Google Patents

Abstract comment abstract generation method based on commodity aspect alignment Download PDF

Info

Publication number
CN111930930A
CN111930930A CN202010663601.6A CN202010663601A CN111930930A CN 111930930 A CN111930930 A CN 111930930A CN 202010663601 A CN202010663601 A CN 202010663601A CN 111930930 A CN111930930 A CN 111930930A
Authority
CN
China
Prior art keywords
comment
abstract
commodity
comments
review
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010663601.6A
Other languages
Chinese (zh)
Other versions
CN111930930B (en
Inventor
潘浩杰
蔡登�
杨荣钦
周鑫
王睿
刘晓钟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN202010663601.6A priority Critical patent/CN111930930B/en
Publication of CN111930930A publication Critical patent/CN111930930A/en
Application granted granted Critical
Publication of CN111930930B publication Critical patent/CN111930930B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/34Browsing; Visualisation therefor
    • G06F16/345Summarisation for human users
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/06Buying, selling or leasing transactions
    • G06Q30/0601Electronic shopping [e-shopping]
    • G06Q30/0641Shopping interfaces
    • G06Q30/0643Graphical representation of items or shoppers

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Linguistics (AREA)
  • Finance (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biophysics (AREA)
  • Health & Medical Sciences (AREA)
  • Accounting & Taxation (AREA)
  • Databases & Information Systems (AREA)
  • Development Economics (AREA)
  • Economics (AREA)
  • Marketing (AREA)
  • Strategic Management (AREA)
  • General Business, Economics & Management (AREA)
  • Machine Translation (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention discloses an abstract review summary generation method based on commodity aspect alignment, which comprises the following steps: (1) the method comprises the steps of obtaining commodity comment data, dividing comments in a commodity, screening out low-quality comments and high-quality comments, and constructing the screened comments into a multi-comment abstract data set in an aspect alignment mode; (2) establishing a sequence model, wherein the sequence model comprises an encoder, a decoder and an attention mechanism based on a recurrent neural network; simultaneously adding an attention mechanism based on aspects; (3) training the sequence model by using the multi-comment abstract data set until the model converges; (4) and performing a comment abstract generation task by using the trained model, and automatically generating an abstract after inputting the comment of the commodity. By utilizing the method, the (review collection, abstract) pairs can be efficiently constructed for neural network model training, and the cost of manual marking is greatly reduced; the trained model can generate high-quality multi-comment abstracts.

Description

Abstract comment abstract generation method based on commodity aspect alignment
Technical Field
The invention belongs to the technical field of natural language processing, and particularly relates to an abstract comment abstract generation method based on commodity aspect alignment.
Background
Review systems are intended to help users make better transactions while shopping online and have become an important part of the active e-commerce environment. However, when the number of comments is large, the user is unlikely to efficiently process the historical comments, considering that most comments may lack key information. On the other hand, the following comments of the e-commerce products are often short, the amount of praise is small, and the commodity Aspect (Aspect) is also small. Therefore, a multi-review summary system is needed to help users efficiently use multiple reviews and digest the most relevant information.
Some of the previous work was extractive, focusing on predicting the overall rating of an entity or estimating the rating of different product features. Abstract methods may be more suitable for summarizing the evaluation text because in the context of a multi-document summary, a decimated review summary may result in a summary that is too verbose or biased towards certain sources. Previous abstract multi-comment summarization work employed an unsupervised approach to reduce multi-comment summaries to a subset of the best phrases to select, and then generate the summary using Natural Language Generation (NLG). The accuracy of training data consisting of manually written review summaries becomes the bottleneck for multiple review summaries. Therefore, we can hardly define a supervised learning paradigm for large-scale evaluation of emerging methods. This has led researchers to employ complex preprocessing methods and predefined rules to summarize reviews.
Modern neural network-based models depend to a large extent on the quality of the available training data, with the most advanced performance already being achieved in abstract text summaries. In particular, this neural network-based model employs an encoder-decoder structure, plus some attention mechanisms. For example, A Neural Attention Model for Abstract Session summary published in 2015 on pages 379 and 389 of The Natural Language Processing Top-level conference, The summary with Point-Generator Networks published in 2017 on The International Top-level computing Neural theory conference for computing Linear concepts was abstracted more simply, and a Pointer mechanism-based generation Model was proposed, so that The task effect is greatly improved.
The method improves the performance of automatic summarization by using data sets such as a Gigaword, New York Time data set, a CNN/Daily Mail corpus and the like. However, the data sets for the multiple review summaries are few and very costly to build. Meanwhile, such models do not consider the most critical part in commodity reviews, namely commodity aspects (Aspect), common aspects such as quality, logistics, customer service, and the like.
Disclosure of Invention
The invention provides an abstract comment abstract generation method based on commodity aspect alignment, which can be used for generating abstract comment abstract under large-scale unsupervised comment linguistic data and generating high-quality multi-comment abstract on the aspects of fluency, diversity, information richness and the like.
An abstract review summary generation method based on commodity aspect alignment comprises the following steps:
(1) the method comprises the steps of obtaining commodity comment data, dividing comments in a commodity, screening out low-quality comments and high-quality comments, and constructing the screened comments into a multi-comment abstract data set in an aspect alignment mode;
(2) establishing a sequence model, wherein the sequence model comprises an encoder, a decoder and an attention mechanism based on a recurrent neural network; simultaneously adding an attention mechanism based on aspects;
(3) training the sequence model by using the multi-comment abstract data set until the model converges;
(4) and performing a comment abstract generation task by using the trained model, and automatically generating an abstract after inputting the comment of the commodity.
In the present invention, the Aspect of a commodity (Aspect) refers to quality, logistics, customer service, and the like involved in the review.
The method comprises the steps of firstly constructing a review set-abstract pair used as end-to-end neural network abstract model training based on aspect alignment based on large-scale unsupervised commodity review corpora, and then constructing a neural network model of an encoder-decoder, an attention mechanism and an aspect attention mechanism, wherein the model can directly generate the review abstract after being trained in the data. The method can reduce the cost of manually marking the end-to-end comment abstract data set, and meanwhile, the trained model can generate high-quality multi-comment abstract on aspects of fluency, diversity, information richness and the like.
The specific process of the step (1) is as follows:
(1-1) collecting comment data of a commodity, pruning some automatically generated low-quality comments (according to certain heuristic rules, the comments are regarded as meaningless comments), and deleting more than 20 times of high-frequency comments displayed under the same product;
(1-2) running an Aspect extractor on these remaining comment data and deleting comments that do not cover any predefined aspects; thereby obtaining reviews of each product containing different aspects; the reviews convey the user's opinion of one or more aspects of the product.
(1-3) for each product, first finding high quality reviews with a number of likes greater than 10 and containing more than 3 facets; then for each corresponding aspect, finding 10-40 low-quality comments to form a low-quality comment set, wherein the low-quality comments include comments with the number of praise less than 1 and only include one aspect;
(1-4) repeating the steps to generate a plurality of groups of (review sets, abstracts) pairs consisting of the low-quality review sets and the corresponding high-quality reviews as review abstract data sets.
In the step (2), the sequence model mainly comprises an encoder and a decoder, and a long-time memory network is used as a main parameter component of the sequence model; the cyclic neural network in the encoder is a bidirectional long-time and short-time memory network.
The sequence model adds an attention mechanism for generating original texts of the words and the comment sets and an attention mechanism for generating words and Aspect vectors.
In the step (2), the working process of the sequence model is as follows:
(2-1) the encoder encodes each word of each comment in the comment set into a vector, and the vector generated for each comment generates a corresponding comment vector through a self-attention mechanism;
(2-2) randomly initializing an aspect vector for each predefined aspect;
(2-3) Co-attentive to the set of aspect vectors of the set of review vectors (Co-Attention);
(2-4) adopting an attention mechanism to carry out weighted summation on the vector of each word of each comment in the comment set, the vector of each comment subjected to common attention and the vector of each aspect subjected to common attention by the hidden layer vector generated by the decoder each time, and integrating to obtain a context vector ctThen the hidden layer vector htAnd a context vector ctAnd integrating by using a linear function and sending into a softmax function to obtain the probability distribution of the prediction sequence.
In step (2-4), the formula of the softmax function is:
P(yt|y<t,x)=softmax(Wpct+Wqht+bp)
wherein, Wp、WqAnd bpAre all parameters to be trained, ytIs the t-th word output by the decoder.
In step (3), an Adam optimizer is used to train a sequence model on the multi-review summary dataset.
Compared with the prior art, the invention has the following beneficial effects:
1. the invention uses a commodity alignment-based method, can efficiently construct (review set, abstract) pairs for neural network model training, greatly reduces the cost of manual labeling, and can generate an abstract multi-review abstract data set with higher quality.
2. The invention fully utilizes commodity information to improve the traditional neural network text abstract model based on an encoder-decoder + attention mechanism, so that the neural network text abstract model is more suitable for the scene of multi-comment abstract, and the trained model can generate high-quality multi-comment abstract on aspects of fluency, diversity, information richness and the like.
Drawings
FIG. 1 is a schematic diagram of the overall structure of a multi-comment summary sequence model based on aspect alignment in the present invention.
Detailed Description
The invention will be described in further detail below with reference to the drawings and examples, which are intended to facilitate the understanding of the invention without limiting it in any way.
As shown in fig. 1, a method for generating abstract review summary based on commodity aspect alignment includes the following steps:
and S01, automatically constructing a large-scale abstract multi-comment summary data set.
This embodiment constructs a multi-review summary dataset starting from 17,936,900,612 product reviews under the category of "clothing".
(1) Reviews of the category that are not related to "clothing" are screened out and some automatically generated reviews are pruned, which are considered meaningless reviews by some heuristic rules (e.g., heuristic rules). More than 20 reviews displayed under the same product are deleted. After screening, a total of 246,585,986 reviews were obtained.
(2) An aspect extractor is run on these comments and comments that do not cover any predefined aspects are removed. At this stage, 153,702,854 reviews have been obtained, accounting for approximately 62.33% of the available reviews, which convey the user's opinion of one or more aspects of the product.
(3) Comments containing aspects have been collected by product ID. For each product, first find "high quality reviews" (> 10 praise, containing > 3 facets), and then for the corresponding facets, find a set of 10-40 (< 1 praise, containing 1 facet) "low quality reviews". Finally 58,418 (panel-summary) pairs were obtained.
S02, training a sequence model of the multi-comment abstract.
Given comment list R ═ { R ═ R(1),r(2),...,r(m)The model of the invention aims to generate an output summary y ═ w1,w2,...,wk}. First, each comment is directly encoded and an unordered set is obtained to represent the comment. Since the reviews represent sequential data with variable length, long-term short-term memory networks are used to learn their feature representation. The use of unordered list coding addresses the differences between the review and article summarization tasks and addresses the importance of aspect alignment between reviews and summaries. After the context is obtained, the summary is generated using a neuro-decoder.
(1) Comment coding
Let R be { R ═ R(1),r(2),...,r(m)As an input comment list, the goal is to encode R as S ═ S(1),s(2),...,s(m)}。
For one having niComments r of individual words(i)It is expressed as a sequence of word vectors, i.e.
Figure BDA0002579513970000051
For a single comment, the comment is coded by using a long and short memory cycle network (LSTM) to obtain the dependency between adjacent words
Figure BDA0002579513970000052
Thus, a sequence of concealment vectors is obtained
Figure BDA0002579513970000061
Then for each H(i)Obtaining corresponding vector s through a layer of structured self-attention mechanism(i)
(2) Aspect alignment
Given the facet list, each facet vector is initialized with the average word vector associated with the facet:
Figure BDA0002579513970000062
in order to take advantage of the link between the different aspects, the
Figure BDA0002579513970000063
Through a layer of fully-connected neural network to obtain a coded hidden aspect vector A ═ a1,a2,...,ak}
Figure BDA0002579513970000064
Here WaIs h-by-d in shapeeH is the size of the hidden state, deIs the embedding size of the word, baIs the bias term.
The ith comment information obtained by performing attention mechanism on A
Figure BDA00025795139700000610
Obtained by the following method:
Figure BDA0002579513970000065
ei,j=cosine(s(i),aj)
thus obtaining a new vector sequence of comments
Figure BDA0002579513970000066
The same attention mechanism canTo obtain a sequence of aspect vectors
Figure BDA0002579513970000067
(3) Nerve abstract generator
The main network of the generator is also a long and short memory neural network LSTM, as follows:
Figure BDA0002579513970000068
concatenating the output of each word of each comment via the encoded LSTM to form a vector sequence:
Figure BDA0002579513970000069
by using
Figure BDA0002579513970000071
To pair
Figure BDA0002579513970000072
Using the attention mechanism, 3 context vectors are obtained and compared with
Figure BDA0002579513970000073
Concatenation may result in a final context vector h for outputout,tAs follows:
Figure BDA0002579513970000074
sending the prediction sequence into a softmax function to obtain probability distribution of the prediction sequence; the formula of the softmax function is as follows:
P(yt|y<t,x)=softmax(Wphout,t+bp)
in the training process, a cross entropy loss function is adopted and optimized by using an Adam optimizer.
S03, evaluating the performance of the sequence model on the data set.
Evaluating the performance of the sequence model by using (1) an abstract generation index ROUGE (2) an Aspect accuracy rate (3) and comparing the sequence model with 3 basic models, wherein the three basic models are respectively as follows:
i) 1, arbitrarily splicing three different low-quality comments;
ii) Seq-to-Seq + attn, which is a simple encoder-decoder + attention model mentioned in the introduction;
iii) Pointer-generator, which is an abstract text abstract generation model published at the front edge of ACL 2017 in background introduction, provides a word of 'copying' input document by a Pointer mechanism, so that the generated abstract has higher quality.
The final comparison is shown in table 1:
TABLE 1
Figure BDA0002579513970000075
Three models were evaluated manually using i) whether it was generated for AI ii) whether it was smooth iii) whether it was useful to understand a commodity, and the results are shown in table 2:
TABLE 2
Figure BDA0002579513970000081
It can be seen that the model of the invention has a very great promotion effect regardless of automatic evaluation or manual evaluation.
The embodiments described above are intended to illustrate the technical solutions and advantages of the present invention, and it should be understood that the above-mentioned embodiments are only specific embodiments of the present invention, and are not intended to limit the present invention, and any modifications, additions and equivalents made within the scope of the principles of the present invention should be included in the scope of the present invention.

Claims (7)

1. An abstract review summary generation method based on commodity aspect alignment is characterized by comprising the following steps:
(1) the method comprises the steps of obtaining commodity comment data, dividing comments in a commodity, screening out low-quality comments and high-quality comments, and constructing the screened comments into a multi-comment abstract data set in an aspect alignment mode;
(2) establishing a sequence model, wherein the sequence model comprises an encoder, a decoder and an attention mechanism based on a recurrent neural network; simultaneously adding an attention mechanism based on aspects;
(3) training the sequence model by using the multi-comment abstract data set until the model converges;
(4) and performing a comment abstract generation task by using the trained model, and automatically generating an abstract after inputting the comment of the commodity.
2. The method for generating abstract review summary based on commodity aspect alignment according to claim 1, wherein the concrete process of step (1) is as follows:
(1-1) collecting comment data of a commodity, pruning some automatically generated low-quality comments, and deleting more than 20 times of high-frequency comments displayed under the same product;
(1-2) running an Aspect extractor on these remaining comment data and deleting comments that do not cover any predefined aspects; thereby obtaining reviews of each product containing different aspects;
(1-3) for each product, first finding high quality reviews with a number of likes greater than 10 and containing more than 3 facets; then for each corresponding aspect, finding 10-40 low-quality comments to form a low-quality comment set, wherein the low-quality comments include comments with the number of praise less than 1 and only include one aspect;
(1-4) repeating the steps to generate a plurality of groups of (review sets, abstracts) pairs consisting of the low-quality review sets and the corresponding high-quality reviews as review abstract data sets.
3. The method for generating abstract review summary based on commodity aspect alignment according to claim 1, wherein in the step (2), the recurrent neural network in the encoder is a bidirectional long-term and short-term memory network.
4. The method for generating abstract review summary based on commodity Aspect alignment of claim 1, wherein in step (2), the sequence model adds an attention mechanism for generating words and original text of the review set and an attention mechanism for generating words and Aspect vectors.
5. The method for generating abstract review summary based on commodity aspect alignment according to claim 1, wherein in the step (2), the sequence model works as follows:
(2-1) the encoder encodes each word of each comment in the comment set into a vector, and the vector generated for each comment generates a corresponding comment vector through a self-attention mechanism;
(2-2) randomly initializing an aspect vector for each predefined aspect;
(2-3) co-attentive to the set of aspect vectors of the set of review vectors;
(2-4) adopting an attention mechanism to carry out weighted summation on the vector of each word of each comment in the comment set, the vector of each comment subjected to common attention and the vector of each aspect subjected to common attention by the hidden layer vector generated by the decoder each time, and integrating to obtain a context vector ctThen the hidden layer vector htAnd a context vector ctAnd integrating by using a linear function and sending into a softmax function to obtain the probability distribution of the prediction sequence.
6. The method for generating abstract review summary based on commodity aspect alignment according to claim 5, wherein in the step (2-4), the formula of the softmax function is:
P(yt|y<t,x)=softmax(Wpct+Wqht+bp)
wherein, Wp、WqAnd bpAre all parameters to be trained, ytIs the t-th word output by the decoder.
7. The method for generating abstract review summary based on commodity aspect alignment according to claim 1, wherein in step (3), an Adam optimizer is used to train a sequence model on the multi-review summary data set.
CN202010663601.6A 2020-07-10 2020-07-10 Abstract comment abstract generation method based on commodity aspect alignment Active CN111930930B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010663601.6A CN111930930B (en) 2020-07-10 2020-07-10 Abstract comment abstract generation method based on commodity aspect alignment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010663601.6A CN111930930B (en) 2020-07-10 2020-07-10 Abstract comment abstract generation method based on commodity aspect alignment

Publications (2)

Publication Number Publication Date
CN111930930A true CN111930930A (en) 2020-11-13
CN111930930B CN111930930B (en) 2022-09-23

Family

ID=73312434

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010663601.6A Active CN111930930B (en) 2020-07-10 2020-07-10 Abstract comment abstract generation method based on commodity aspect alignment

Country Status (1)

Country Link
CN (1) CN111930930B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112541343A (en) * 2020-12-03 2021-03-23 昆明理工大学 Semi-supervised counterstudy cross-language abstract generation method based on word alignment

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109800390A (en) * 2018-12-21 2019-05-24 北京石油化工学院 A kind of calculation method and device of individualized emotion abstract

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109800390A (en) * 2018-12-21 2019-05-24 北京石油化工学院 A kind of calculation method and device of individualized emotion abstract

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
余传明等: "面向查询的观点摘要模型研究:以Debatepedia为数据源", 《情报学报》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112541343A (en) * 2020-12-03 2021-03-23 昆明理工大学 Semi-supervised counterstudy cross-language abstract generation method based on word alignment
CN112541343B (en) * 2020-12-03 2022-06-14 昆明理工大学 Semi-supervised counterstudy cross-language abstract generation method based on word alignment

Also Published As

Publication number Publication date
CN111930930B (en) 2022-09-23

Similar Documents

Publication Publication Date Title
Shrestha et al. Deep learning sentiment analysis of amazon. com reviews and ratings
CN109325112A (en) A kind of across language sentiment analysis method and apparatus based on emoji
CN105183833A (en) User model based microblogging text recommendation method and recommendation apparatus thereof
CN107341145A (en) A kind of user feeling analysis method based on deep learning
CN111178053B (en) Text generation method for generating abstract extraction by combining semantics and text structure
CN111079409A (en) Emotion classification method by using context and aspect memory information
Zhu et al. Prompt-learning for short text classification
CN112016002A (en) Mixed recommendation method integrating comment text level attention and time factors
CN113821635A (en) Text abstract generation method and system for financial field
Upadhya et al. Deep neural network models for question classification in community question-answering forums
CN113934835B (en) Retrieval type reply dialogue method and system combining keywords and semantic understanding representation
CN111930930B (en) Abstract comment abstract generation method based on commodity aspect alignment
CN113220964B (en) Viewpoint mining method based on short text in network message field
CN109241272A (en) A kind of Chinese text abstraction generating method, computer-readable storage media and computer equipment
Guo et al. Multimodal news recommendation based on deep reinforcement learning
Wu et al. PESI: Personalized Explanation recommendation with Sentiment Inconsistency between ratings and reviews
Yu et al. Multi-module Fusion Relevance Attention Network for Multi-label Text Classification.
Chen et al. Understanding emojis for financial sentiment analysis
CN114022233A (en) Novel commodity recommendation method
Raja et al. Deep learning-based sentiment analysis of Trip Advisor reviews
Singh Stockgram: deep learning model for digitizing financial communications via Natural Language generation
CN117951390B (en) Personalized content recommendation method and system based on large language model
Shiri et al. Meme it Up: Patterns of Emoji Usage on Twitter
Ha et al. Unsupervised Sentence Embeddings for Answer Summarization in Non-factoid CQA
Feng et al. Negative Examples Sampling Based on Factorization Machines for OCCF

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant