CN111639969B - Dynamic incentive calculation method, system, equipment and medium for crowdsourcing system - Google Patents
Dynamic incentive calculation method, system, equipment and medium for crowdsourcing system Download PDFInfo
- Publication number
- CN111639969B CN111639969B CN202010466953.2A CN202010466953A CN111639969B CN 111639969 B CN111639969 B CN 111639969B CN 202010466953 A CN202010466953 A CN 202010466953A CN 111639969 B CN111639969 B CN 111639969B
- Authority
- CN
- China
- Prior art keywords
- neural network
- cyclic neural
- user
- network model
- task
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000004364 calculation method Methods 0.000 title abstract description 15
- 238000003062 neural network model Methods 0.000 claims abstract description 40
- 125000004122 cyclic group Chemical group 0.000 claims abstract description 35
- 230000008901 benefit Effects 0.000 claims abstract description 30
- 238000000034 method Methods 0.000 claims abstract description 25
- 238000012549 training Methods 0.000 claims abstract description 20
- 230000005284 excitation Effects 0.000 claims abstract 16
- 230000006870 function Effects 0.000 claims description 20
- 238000013528 artificial neural network Methods 0.000 claims description 15
- 238000010276 construction Methods 0.000 claims description 4
- 230000005540 biological transmission Effects 0.000 claims description 2
- 238000004590 computer program Methods 0.000 claims description 2
- 238000013138 pruning Methods 0.000 claims description 2
- 238000013441 quality evaluation Methods 0.000 claims 1
- 238000004422 calculation algorithm Methods 0.000 abstract description 5
- 238000002474 experimental method Methods 0.000 abstract description 2
- 238000004088 simulation Methods 0.000 abstract description 2
- 230000000306 recurrent effect Effects 0.000 description 17
- 230000006872 improvement Effects 0.000 description 4
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000010845 search algorithm Methods 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 230000006399 behavior Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000001303 quality assessment method Methods 0.000 description 1
- 239000013598 vector Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0201—Market modelling; Market analysis; Collecting market data
- G06Q30/0206—Price or cost determination based on market factors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Business, Economics & Management (AREA)
- General Physics & Mathematics (AREA)
- Strategic Management (AREA)
- Finance (AREA)
- Development Economics (AREA)
- Accounting & Taxation (AREA)
- Data Mining & Analysis (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Evolutionary Computation (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Artificial Intelligence (AREA)
- Entrepreneurship & Innovation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Game Theory and Decision Science (AREA)
- Economics (AREA)
- Marketing (AREA)
- General Business, Economics & Management (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
Description
技术领域technical field
本发明涉及数据质量提升问题,特别是指一种用于众包系统的动态激励计算方法、系统、设备及介质。The present invention relates to the problem of data quality improvement, in particular to a dynamic incentive calculation method, system, equipment and medium for a crowdsourcing system.
背景技术Background technique
在实际应用中,存在着很多人类可以轻易完成但机器难以直接完成的问题。例如给出两张不同清晰度的图片,人类可以快速准确分辨但机器却难以识别。同样的例子还有自然语言情感清晰的判断。在这样的背景下,众包平台得到了广泛的关注和发展。In practical applications, there are many problems that humans can easily complete but machines cannot directly complete. For example, given two pictures of different resolutions, humans can quickly and accurately distinguish them, but it is difficult for machines to recognize them. The same example is the clear judgment of natural language emotion. In this context, crowdsourcing platforms have received extensive attention and development.
众包平台指的是一个网络工作分配平台,需求者在平台上发布各种任务,用户浏览并选择任务,需求者根据用户提交的工作质量来给予一定的奖励。Crowdsourcing platform refers to a network job distribution platform. Demanders publish various tasks on the platform, users browse and select tasks, and demanders give certain rewards according to the quality of work submitted by users.
随着众包平台的不断发展,如何给予激励以提升问答数据质量成为了一个关键问题。有研究发现,适当的激励(如声望、金钱等)可以提高问答质量,进而提升需求者的最终受益。然而,过多地给予激励值会使得需求者的总收益下降,进而使得答案数据的总体质量下降,过少的激励值会打击用户回复的积极性,任务可能无法完成。在这样的情况下,良好的激励机制不仅要求合理地激励用户以获取高质量回复,同时也要使得需求者可以通过用户的答案获得相对较高的收益。With the continuous development of crowdsourcing platforms, how to give incentives to improve the quality of question answering data has become a key issue. Some studies have found that appropriate incentives (such as prestige, money, etc.) can improve the quality of questions and answers, thereby increasing the ultimate benefit of the demander. However, giving too much incentive value will reduce the total income of the demander, which in turn will reduce the overall quality of the answer data, and too little incentive value will discourage the enthusiasm of users to reply, and the task may not be completed. Under such circumstances, a good incentive mechanism not only requires reasonable incentives for users to obtain high-quality replies, but also enables demanders to obtain relatively high returns through users' answers.
针对众包平台上的激励计算问题,国内外学者已经做出了一些工作,但这些工作还存在局限性:(1)激励模式单一,带来的整体数据质量提升效果均有限;(2)每个完成任务的用户获得相同利益,没有考虑到答案质量、用户行为等影响。Scholars at home and abroad have done some work on the incentive calculation problem on the crowdsourcing platform, but these works still have limitations: (1) the incentive model is single, and the overall data quality improvement effect is limited; (2) every Each user who completes the task gets the same benefits, without considering the impact of answer quality and user behavior.
发明内容Contents of the invention
本发明实施例提供一种用于众包系统的动态激励计算方法、系统、设备及介质,以解决传统方案单一且未考虑答案质量的问题,根据回答者的以往任务问答历史数据来决定是否给予回答者一定的激励值以最大化任务数据的质量。Embodiments of the present invention provide a dynamic incentive calculation method, system, device, and medium for a crowdsourcing system to solve the problem that the traditional solution is single and does not consider the quality of the answer. Respondents are motivated to maximize the quality of task data.
为了达到上述目的,本发明采用如下技术方案:In order to achieve the above object, the present invention adopts following technical scheme:
第一方面,本发明实施例提供一种用于众包系统的动态激励计算方法,该方法包括如下步骤:In the first aspect, an embodiment of the present invention provides a dynamic incentive calculation method for a crowdsourcing system, the method includes the following steps:
获取需求者在众包平台上的任务数据以及参与用户的历史任务问答数据;Obtain the task data of demanders on the crowdsourcing platform and the historical task question and answer data of participating users;
将任务分配给参与用户;Assign tasks to participating users;
针对每个参与用户构建一个循环神经网络模型;Construct a recurrent neural network model for each participating user;
根据参与用户的历史任务问答数据,训练循环神经网络模型;According to the historical task question and answer data of participating users, train the recurrent neural network model;
依据参与用户、任务以及循环神经网络模型的预测结果,计算不同激励值所带来的最终收益大小判断是否给予当前用户激励值;Based on the prediction results of the participating users, tasks and the cyclic neural network model, calculate the final benefits brought by different incentive values and judge whether to give the current user an incentive value;
收集所有参与用户的答案,以得到需求者所需的结果。Collect answers from all participating users to get the results required by demanders.
进一步地,获取的任务数据包括:任务个数、每轮分配任务数、需要问答的任务的总集合。而参与用户的历史任务问答数据包含用户以往任务问答数目和质量。Further, the acquired task data includes: the number of tasks, the number of tasks assigned in each round, and the total set of tasks requiring question and answer. The historical task question and answer data of participating users includes the number and quality of the user's past task questions and answers.
进一步地,所述循环神经网络模型是一种时间序列模型,用于预测在给定激励值时用户的输出答案质量水平,循环神经网络模型其由多个全连接层组成,在每个时间节点接收上一个时间节点的输出,构成一个循环的神经网络。Further, the recurrent neural network model is a time series model, which is used to predict the quality level of the user's output answer when the incentive value is given. The recurrent neural network model is composed of multiple fully connected layers, and at each time node Receive the output of the previous time node to form a cyclic neural network.
进一步地,所述循环神经网络模型主要的参数如下:Further, the main parameters of the recurrent neural network model are as follows:
1)需求者可以决定在每个时间节点t是否给予用户激励值,记作at,at为1则给予激励值,0则为否;1) The demander can decide whether to give the user an incentive value at each time node t, denoted as a t , if a t is 1, the incentive value is given, and if 0 is no;
2)循环神经网络模型的输出为变量yt,代表用户完成任务为高质量的概率,yt越接近1则答案的质量越高,越接近0则答案的质量越低;2) The output of the cyclic neural network model is the variable y t , which represents the probability that the user completes the task with high quality. The closer y t is to 1, the higher the quality of the answer is, and the closer y t is to 0, the lower the quality of the answer;
3)存在多个隐藏状态,神经网络中输入与隐藏状态、隐藏状态之间、隐藏状态和输出之间的传递参数通过训练循环神经网络模型得到。3) There are multiple hidden states, and the transfer parameters between the input and hidden states, between hidden states, and between hidden states and outputs in the neural network are obtained by training the cyclic neural network model.
进一步地,所述循环神经网络模型的构建步骤如下:Further, the construction steps of the recurrent neural network model are as follows:
训练数据集为用户的历史任务问答数据序列<at,yt>,通过循环神经网络的反向传播更新并优化模型参数,使其在训练数据集上的表现得到优化。经过多次迭代训练后即可得到所需的答案质量评估模型。The training data set is the user's historical task question answering data sequence <a t , y t > , and the model parameters are updated and optimized through the backpropagation of the recurrent neural network to optimize its performance on the training data set. After multiple iterations of training, the desired answer quality assessment model can be obtained.
进一步地,依据参与用户、任务以及循环神经网络的预测结果,计算不同激励值所带来的最终受益大小判断是否给予当前用户激励值,包括:Further, according to the prediction results of participating users, tasks and cyclic neural network, calculate the final benefit brought by different incentive values to judge whether to give the current user an incentive value, including:
对于某一用户,给定相应的激励值,作为循环神经网络模型的输入,获得激励值相对应的预测答案质量;For a certain user, given the corresponding incentive value, as the input of the recurrent neural network model, the predicted answer quality corresponding to the incentive value is obtained;
通过预测的答案质量构建最终收益函数,最终收益函数包含完成下一个任务的收益和完成未来任务所获得的收益;Construct the final reward function through the predicted answer quality, and the final reward function includes the reward for completing the next task and the reward for completing future tasks;
求解相应的最终收益函数,即可得到相应激励值所获得的最终受益,由此决定是否给予当前用户激励值。Solving the corresponding final benefit function can obtain the final benefit obtained by the corresponding incentive value, so as to decide whether to give the current user an incentive value.
进一步地,所述求解相应的最终收益函数采用启发式剪枝策略,降低模型计算复杂度。Further, the solution to the corresponding final revenue function adopts a heuristic pruning strategy to reduce the computational complexity of the model.
第二方面,本发明实施例还提供一种用于众包系统的动态激励计算系统,包括:In the second aspect, the embodiment of the present invention also provides a dynamic incentive computing system for a crowdsourcing system, including:
获取模块,用于获取需求者在众包平台上的任务数据以及参与用户的历史任务问答数据;The acquisition module is used to acquire the task data of demanders on the crowdsourcing platform and the historical task question and answer data of participating users;
分配模块,用于将任务分配给参与用户;Assignment module for assigning tasks to participating users;
模型构建模块,用于针对每个参与用户构建一个循环神经网络模型;A model building block for building a recurrent neural network model for each participating user;
模型训练模块,用于根据参与用户的历史任务问答数据,训练循环神经网络模型;The model training module is used to train the recurrent neural network model according to the historical task question and answer data of the participating users;
决策模块,用于依据参与用户、任务以及循环神经网络模型的预测结果,计算不同激励值所带来的最终收益大小判断是否给予当前用户激励值;The decision-making module is used to calculate the final benefits brought by different incentive values according to the prediction results of participating users, tasks and the cyclic neural network model, and judge whether to give the current user an incentive value;
结果输出模块,用于收集所有参与用户的答案,以得到需求者所需的结果。The result output module is used to collect the answers of all participating users to obtain the results required by the demanders.
第三方面,本发明实施例还提供一种电子设备,包括:In a third aspect, the embodiment of the present invention also provides an electronic device, including:
一个或多个处理器;one or more processors;
存储器,用于存储一个或多个程序;memory for storing one or more programs;
当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现如第一方面所述的方法。When the one or more programs are executed by the one or more processors, the one or more processors implement the method as described in the first aspect.
第四方面,本发明实施例还提供一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现如第一方面所述的方法。In a fourth aspect, an embodiment of the present invention further provides a computer-readable storage medium, on which a computer program is stored, and when the program is executed by a processor, the method described in the first aspect is implemented.
根据以上技术方案,该方法包括:将任务分派给用户,通过用户的历史任务问答数据建立一个循环神经网络模型用于预测;通过训练好的模型,本方法预测是否给予激励值的用户问答数据质量,并由此计算最终的收益。本方法模型不仅仅考虑到下一个任务的收益,更考虑了未来所有任务完成对最终收益产生的影响,同时可以动态地决定当前用户接受任务时是否给予激励值,以最大化需求者所获得的最终受益。模拟实验证实了本发明在复杂情况下的高效性和鲁棒性。众包平台上的实际实验也显示了本发明相对于传统方法的高效性及优越性。According to the above technical solution, the method includes: assigning the task to the user, and establishing a recurrent neural network model for prediction through the user's historical task question and answer data; through the trained model, this method predicts whether to give the user the quality of the user's question and answer data with an incentive value , and calculate the final income from it. This method model not only considers the income of the next task, but also considers the impact of the completion of all future tasks on the final income. At the same time, it can dynamically decide whether to give the incentive value when the current user accepts the task, so as to maximize the demander’s income. benefit in the end. Simulation experiments confirm the efficiency and robustness of the invention in complex situations. The actual experiment on the crowdsourcing platform also shows the high efficiency and superiority of the present invention compared with the traditional method.
附图说明Description of drawings
此处所说明的附图用来提供对本发明的进一步理解,构成本发明的一部分,本发明的示意性实施例及其说明用于解释本发明,并不构成对本发明的不当限定。在附图中:The accompanying drawings described here are used to provide a further understanding of the present invention, and constitute a part of the present invention. The schematic embodiments of the present invention and their descriptions are used to explain the present invention, and do not constitute improper limitations to the present invention. In the attached picture:
图1是本发明实施例的一种用于众包系统的动态激励计算方法的流程图;Fig. 1 is a flow chart of a dynamic incentive calculation method for a crowdsourcing system according to an embodiment of the present invention;
图2是本发明实施例中的模型系统框图;Fig. 2 is the model system block diagram in the embodiment of the present invention;
图3是本发明实施例中循环神经网络(RNN)的示意图;Fig. 3 is the schematic diagram of recurrent neural network (RNN) in the embodiment of the present invention;
图4是本发明实施例的一种用于众包系统的动态激励计算系统的框图。Fig. 4 is a block diagram of a dynamic incentive computing system used in a crowdsourcing system according to an embodiment of the present invention.
具体实施方式Detailed ways
为了使本发明的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本发明进行进一步详细说明。应该理解,此处所描述的具体实施例仅用以解释本发明,并不用于限定本发明。In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention.
相反,本发明涵盖任何由权利要求定义的在本发明的精髓和范围上做的替代、修改、等效方法以及方案。进一步,为了使公众对本发明有更好的了解,在下文对本发明的细节描述中,详尽描述了一些特定的细节部分。On the contrary, the invention covers any alternatives, modifications, equivalent methods and schemes within the spirit and scope of the invention as defined by the claims. Further, in order to make the public have a better understanding of the present invention, some specific details are described in detail in the detailed description of the present invention below.
实施例一Embodiment one
图1是本发明实施例的一种用于众包系统的动态激励计算方法的流程图,如图1所示,该方法包括以下步骤:Fig. 1 is a flow chart of a dynamic incentive calculation method for a crowdsourcing system according to an embodiment of the present invention. As shown in Fig. 1, the method includes the following steps:
步骤S100:获取需求者在众包平台上所提交的任务数据以及参与用户的历史任务问答数据。任务数据包括需要完成的任务集合O,任务总数C,每轮所需要完成的任务数目c。Step S100: Obtain task data submitted by demanders on the crowdsourcing platform and historical task question and answer data of participating users. The task data includes the task set O that needs to be completed, the total number of tasks C, and the number c of tasks that need to be completed in each round.
步骤S200:根据如上的任务信息,将每轮的任务分配给用户。Step S200: According to the above task information, assign the tasks of each round to the users.
步骤S300:根据参与用户的历史任务问答数据,对每个参与用户构建一个循环神经网络模型用于预测。Step S300: Construct a recurrent neural network model for each participating user for prediction according to the historical task question and answer data of the participating users.
步骤S301:针对每一个用户,建立相应的循环神经网络RNN模型,所述RNN模型是一种时间序列模型,用于预测在给定激励值时用户的输出答案质量水平。RNN模型其由多个全连接层组成,在每个时间节点接收上一个时间节点的输出,构成一个循环神经网络。其主要的参数如下:Step S301: For each user, establish a corresponding recurrent neural network RNN model, the RNN model is a time series model, used to predict the quality level of the user's output answer when the incentive value is given. The RNN model is composed of multiple fully connected layers, and receives the output of the previous time node at each time node to form a recurrent neural network. Its main parameters are as follows:
1)需求者可以决定在每个节点t是否给予用户激励值,记作at。at为1则给予激励值,0则为否。1) The demander can decide whether to give the user an incentive value at each node t, denoted as a t . When a t is 1, an incentive value is given, and 0 means no.
2)循环神经网络模型的输出为变量yt,代表用户完成任务为高质量的概率,yt越接近1则答案的质量越高,越接近0则答案的质量越低;2) The output of the cyclic neural network model is the variable y t , which represents the probability that the user completes the task with high quality. The closer y t is to 1, the higher the quality of the answer is, and the closer y t is to 0, the lower the quality of the answer;
3)存在多个隐藏状态。神经网络中输入与隐藏状态、隐藏状态之间、隐藏状态和输出之间的传递参数U,W,V可以通过训练循环神经网络模型得到。3) There are multiple hidden states. The transmission parameters U, W, and V between the input and the hidden state, between the hidden states, and between the hidden state and the output in the neural network can be obtained by training the cyclic neural network model.
步骤S400:根据参与用户的历史回答数据,训练循环神经网络模型。Step S400: Train the cyclic neural network model according to the historical answer data of the participating users.
步骤S401:训练循环神经网络模型的步骤如下:训练数据集为用户的历史任务问答数据序列<at,yt>,通过循环神经网络的反向传播,模型的传递参数得到修改,在训练数据集上的表现得到优化。经过多次训练后即可得到所需的答案质量预测模型RNN。Step S401: The steps of training the cyclic neural network model are as follows: the training data set is the user's historical task question answering data sequence <a t , y t > , through the backpropagation of the cyclic neural network, the transfer parameters of the model are modified, and the training data The performance on the set is optimized. After many times of training, the required answer quality prediction model RNN can be obtained.
步骤S500:依据参与用户、任务以及循环神经网络模型的预测结果,计算不同激励值所带来的最终受益大小判断是否给予当前用户激励值。Step S500: According to the prediction results of participating users, tasks and the cyclic neural network model, calculate the final benefit brought by different incentive values and judge whether to give the current user an incentive value.
步骤S501:首先,通过RNN模型获得相应的预测结果。在训练集上训练好的循环神经网络RNN可表示为:Step S501: First, obtain corresponding prediction results through the RNN model. The cyclic neural network RNN trained on the training set can be expressed as:
fθ:a→y (1)f θ : a→y (1)
其中a和y均为L维向量,代表L个任务是否给予激励值以及答案为高质量回复的概率。其中前L-1维数据为用户之前问答的数据,定义为当前状态s,aL为当前任务的输入。于是RNN模型也可以记作y(s,a),s为当前状态,a为当前任务是否给予激励值,其输出为该状态和激励值水平下当前用户有高质量答案的概率。Where a and y are both L-dimensional vectors, representing whether the L tasks are given incentive values and the probability that the answer is a high-quality reply. Among them, the first L-1 dimensional data is the data of the user’s previous question and answer, which is defined as the current state s, and a L is the input of the current task. Therefore, the RNN model can also be recorded as y(s, a), s is the current state, a is whether the current task is given an incentive value, and its output is the probability that the current user has a high-quality answer under the state and incentive value level.
步骤S502:其次,通过预测的结果构建相应的损失函数。本方法通过计算需求者的在线收益期望来动态的决定是否给予激励值。假设某一用户已完成了数次任务并且还有数个未完成的任务,总收益的效用函数不仅仅只包含完成下一个任务的收益,同时也考虑到完成未来任务所获得的收益。在已给当前状态s,未来需要完成的工作数量tn和当前是否给激励值a时,收益函数记作E[U(s;a;tn)]或记作E[U],定义如下:Step S502: Second, construct a corresponding loss function based on the predicted result. This method dynamically decides whether to give incentive value by calculating the demander's online income expectation. Assuming that a user has completed several tasks and has several unfinished tasks, the utility function of the total revenue not only includes the revenue from completing the next task, but also considers the revenue from completing future tasks. When the current state s is given, the amount of work to be completed in the future t n and whether the current incentive value a is given, the income function is recorded as E[U(s; a; t n )] or E[U], defined as follows :
其中in
F(s,a)=[1-y(s,a)]wl+y(s,a)[wh-I(a)b] (3)F(s,a)=[1-y(s,a)]w l +y(s,a)[w h -I(a)b] (3)
G(s′a,y,tn-1)=maxa′∈{0,1}E[U(s′a,y;a′;tn-1)] (4)G(s' a, y , t n -1) = max a' ∈ {0, 1} E[U(s' a, y ; a'; t n -1)] (4)
I(a)为1当且仅当a=1,其他情况为0。I(a) is 1 if and only if a=1, and 0 otherwise.
步骤S503:通过最大化相应的收益函数决定是否给予激励,从而实现决策。有了如上的收益函数,我们就可以对决策问题进行公式化定义,也就是已知当前状态s,未来需要完成的工作数量tn时求得:Step S503: Decide whether to give incentives by maximizing the corresponding revenue function, so as to realize the decision. With the above income function, we can formulate the definition of the decision-making problem, that is, when the current state s is known and the number of tasks to be completed in the future t n is obtained:
a=arg maxa∈{0,1}E[U(s;a;tn)] (5)a = arg max a ∈ {0, 1} E[U(s; a; t n )] (5)
a=1时则给用户激励值,若为0则不给。When a=1, the incentive value will be given to the user, if it is 0, it will not be given.
收益函数计算问题的动态求解可直接由动态规划求解,但其时间复杂度为指数级。本方案提出了一种高效的启发式算法来解决上述问题。The dynamic solution of the revenue function calculation problem can be solved directly by dynamic programming, but its time complexity is exponential. This scheme proposes an efficient heuristic algorithm to solve the above problems.
1)Beam-Search算法:在某一时刻,所要完成的任务数量可能有一定数量,因此在求此刻的最终收益时时可能需要迭代过多次,导致过高的计算复杂度。因此对于当前时刻的某个任务,无论其是否给予激励值,在计算收益函数的时候我们只考虑之前m个最大收益的影响,其中m为搜索宽度,为人为确定的参数。其他部分的收益忽略。Beam-Search与贪心算法有略微不同之处,该算法考虑前m个最大收益的影响,而贪心算法只考虑之前最大的收益影响。有了之前循环神经网络模型,Beam-Search算法就可以根据用户是否给予激励值来计算不同的收益,从而得出决策。1) Beam-Search algorithm: At a certain moment, there may be a certain number of tasks to be completed, so it may be necessary to iterate too many times when seeking the final income at this moment, resulting in excessive computational complexity. Therefore, for a certain task at the current moment, no matter whether it is given an incentive value or not, we only consider the influence of the previous m maximum revenues when calculating the revenue function, where m is the search width and is an artificially determined parameter. Other parts of the income are ignored. Beam-Search is slightly different from the greedy algorithm. The algorithm considers the influence of the first m largest returns, while the greedy algorithm only considers the previous largest return influence. With the previous cyclic neural network model, the Beam-Search algorithm can calculate different benefits according to whether the user gives the incentive value, so as to make a decision.
步骤S600:收集之前任务的答案,并通过所有用户的答案得到结果。Step S600: Collect the answers of the previous tasks, and obtain the results through the answers of all users.
实施例二Embodiment two
本发明还提供一种用于众包系统的动态激励计算系统的实施例,由于本发明提供的用于众包系统的动态激励计算系统与前述用于众包系统的动态激励计算方法例相对应,该用于众包系统的动态激励计算系统可以通过执行上述方法具体实施方式中的流程步骤来实现本发明的目的,因此用于众包系统的动态激励计算方法的实施例中的解释说明,也适用于动态激励计算系统的实施例,在本发明以下的实施例中将不再赘述。The present invention also provides an embodiment of a dynamic incentive calculation system for a crowdsourcing system, since the dynamic incentive calculation system for a crowdsourcing system provided by the present invention corresponds to the aforementioned example of a dynamic incentive calculation method for a crowdsourcing system , the dynamic incentive calculation system for the crowdsourcing system can realize the purpose of the present invention by executing the process steps in the specific implementation of the above method, so the explanation in the embodiment of the dynamic incentive calculation method for the crowdsourcing system, It is also applicable to the embodiments of the dynamic incentive computing system, which will not be repeated in the following embodiments of the present invention.
如图4,本实施例提供一种用于众包系统的动态激励计算系统,包括:As shown in Figure 4, this embodiment provides a dynamic incentive computing system for crowdsourcing systems, including:
获取模块91,用于获取需求者在众包平台上的任务数据以及参与用户的历史任务问答数据;The obtaining module 91 is used to obtain the task data of the requester on the crowdsourcing platform and the historical task question and answer data of the participating users;
分配模块92,用于将任务分配给参与用户;Assignment module 92, for assigning tasks to participating users;
模型构建模块93,用于针对每个参与用户构建一个循环神经网络模型;Model construction module 93, is used for constructing a recurrent neural network model for each participating user;
模型训练模块94,用于根据参与用户的历史任务问答数据,训练循环神经网络模型;The model training module 94 is used to train the recurrent neural network model according to the historical task question and answer data of the participating users;
决策模块95,用于依据参与用户、任务以及循环神经网络模型的预测结果,计算不同激励值所带来的最终收益大小判断是否给予当前用户激励值;The decision-making module 95 is used to calculate the final profit brought by different incentive values according to the prediction results of the participating users, tasks and the cyclic neural network model, and judge whether to give the current user an incentive value;
结果输出模块96,用于收集所有参与用户的答案,以得到需求者所需的结果。The result output module 96 is used to collect the answers of all participating users, so as to obtain the results required by the demanders.
上述本发明实施例序号仅仅为了描述,不代表实施例的优劣。The serial numbers of the above embodiments of the present invention are for description only, and do not represent the advantages and disadvantages of the embodiments.
在本发明的上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述的部分,可以参见其他实施例的相关描述。In the above-mentioned embodiments of the present invention, the descriptions of each embodiment have their own emphases, and for parts not described in detail in a certain embodiment, reference may be made to relevant descriptions of other embodiments.
在本申请所提供的几个实施例中,应该理解到,所揭露的技术内容,可通过其它的方式实现。其中,以上所描述的设备实施例仅仅是示意性的,例如所述单元的划分,可以为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,单元或模块的间接耦合或通信连接,可以是电性或其它的形式。In the several embodiments provided in this application, it should be understood that the disclosed technical content can be realized in other ways. Wherein, the device embodiments described above are only illustrative. For example, the division of the units may be a logical function division. In actual implementation, there may be other division methods. For example, multiple units or components may be combined or may be Integrate into another system, or some features may be ignored, or not implemented. In another point, the mutual coupling or direct coupling or communication connection shown or discussed may be through some interfaces, and the indirect coupling or communication connection of units or modules may be in electrical or other forms.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or may be distributed to multiple units. Part or all of the units can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
另外,在本发明各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, each functional unit in each embodiment of the present invention may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit. The above-mentioned integrated units can be implemented in the form of hardware or in the form of software functional units.
所述集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本发明的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可为个人计算机、服务器或者网络设备等)执行本发明各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、移动硬盘、磁碟或者光盘等各种可以存储程序代码的介质。If the integrated unit is realized in the form of a software function unit and sold or used as an independent product, it can be stored in a computer-readable storage medium. Based on such an understanding, the essence of the technical solution of the present invention or the part that contributes to the prior art or all or part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium , including several instructions to make a computer device (which may be a personal computer, a server, or a network device, etc.) execute all or part of the steps of the methods described in various embodiments of the present invention. The aforementioned storage media include: U disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), mobile hard disk, magnetic disk or optical disk and other media that can store program codes. .
以上所述是本发明的优选实施方式,应当指出,对于本技术领域的普通技术人员来说,在不脱离本发明所述原理的前提下,还可以做出若干改进和润饰,这些改进和润饰也应视为本发明的保护范围。The above description is a preferred embodiment of the present invention, it should be pointed out that for those of ordinary skill in the art, without departing from the principle of the present invention, some improvements and modifications can also be made, these improvements and modifications It should also be regarded as the protection scope of the present invention.
Claims (9)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010466953.2A CN111639969B (en) | 2020-05-28 | 2020-05-28 | Dynamic incentive calculation method, system, equipment and medium for crowdsourcing system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010466953.2A CN111639969B (en) | 2020-05-28 | 2020-05-28 | Dynamic incentive calculation method, system, equipment and medium for crowdsourcing system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111639969A CN111639969A (en) | 2020-09-08 |
CN111639969B true CN111639969B (en) | 2023-05-26 |
Family
ID=72331212
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010466953.2A Active CN111639969B (en) | 2020-05-28 | 2020-05-28 | Dynamic incentive calculation method, system, equipment and medium for crowdsourcing system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111639969B (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112598286B (en) * | 2020-12-23 | 2024-11-08 | 北京百舸飞驰科技有限公司 | A method, device and electronic device for detecting cheating behavior of crowdsourcing users |
CN113379392A (en) * | 2021-06-29 | 2021-09-10 | 中国科学技术大学 | Method for acquiring high-quality data for numerical tasks in crowdsourcing scene |
CN113393056B (en) * | 2021-07-08 | 2022-11-25 | 山东大学 | Crowdsourcing service supply and demand gap prediction method and system based on time sequence |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140180779A1 (en) * | 2012-12-20 | 2014-06-26 | International Business Machines Corporation | Automated incentive computation in crowdsourcing systems |
US9911088B2 (en) * | 2014-05-01 | 2018-03-06 | Microsoft Technology Licensing, Llc | Optimizing task recommendations in context-aware mobile crowdsourcing |
US10467541B2 (en) * | 2016-07-27 | 2019-11-05 | Intuit Inc. | Method and system for improving content searching in a question and answer customer support system by using a crowd-machine learning hybrid predictive model |
CN108596335B (en) * | 2018-04-20 | 2020-04-17 | 浙江大学 | Self-adaptive crowdsourcing method based on deep reinforcement learning |
CN109889388B (en) * | 2019-03-12 | 2022-04-01 | 湖北工业大学 | Design method of dynamic contract incentive mechanism for mobile crowdsourcing network based on reputation theory |
CN110264272A (en) * | 2019-06-21 | 2019-09-20 | 山东师范大学 | A kind of mobile Internet labor service crowdsourcing platform task optimal pricing prediction technique, apparatus and system |
-
2020
- 2020-05-28 CN CN202010466953.2A patent/CN111639969B/en active Active
Also Published As
Publication number | Publication date |
---|---|
CN111639969A (en) | 2020-09-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Liu et al. | Competing bandits in matching markets | |
CN111639969B (en) | Dynamic incentive calculation method, system, equipment and medium for crowdsourcing system | |
CN111241407B (en) | Personalized search method based on reinforcement learning | |
US8112369B2 (en) | Methods and systems of adaptive coalition of cognitive agents | |
CN114139637B (en) | Multi-agent information fusion method and device, electronic equipment and readable storage medium | |
CN108038622B (en) | Method for recommending users by crowd sensing system | |
Runarsson | Constrained evolutionary optimization by approximate ranking and surrogate models | |
CN114169412B (en) | Federal learning model training method for large-scale industry chain privacy calculation | |
CN111581516A (en) | Investment product recommendation method and related device | |
CN111210072B (en) | Prediction model training and user resource limit determining method and device | |
CN107437111A (en) | Data processing method, medium, device and computing device based on neutral net | |
Kim et al. | Optimizing large-scale fleet management on a road network using multi-agent deep reinforcement learning with graph neural network | |
CN111027709B (en) | Information recommendation method and device, server and storage medium | |
Amini et al. | A BOA-based adaptive strategy with multi-party perspective for automated multilateral negotiations | |
CN108198018A (en) | A kind of trade network system and its implementation for block chain address ranking | |
Lin et al. | Bandit-NAS: Bandit sampling and training method for Neural Architecture Search | |
Huang et al. | An online inference-aided incentive framework for information elicitation without verification | |
Oderanti et al. | Application of strategic fuzzy games to wage increase negotiation and decision problems | |
Chen et al. | A pricing approach toward incentive mechanisms for participant mobile crowdsensing in edge computing | |
CN117669993A (en) | Progressive charging facility planning method, progressive charging facility planning device, terminal and storage medium | |
CN114548409B (en) | Game method and device for unmanned vehicle task assignment based on situation field | |
CN115081253A (en) | Data processing method and device | |
CN114528992A (en) | Block chain-based e-commerce business analysis model training method | |
Abd Elfattah et al. | Carbon price time series forecasting utilizing an optimized ANFIS model | |
Zheng et al. | Making the most cost-effective decision in online paid Q&A community: An expert recommender system with motivation modeling and knowledge pricing |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |