CN114282645A - DQN-based space-time crowdsourcing task allocation method - Google Patents

DQN-based space-time crowdsourcing task allocation method Download PDF

Info

Publication number
CN114282645A
CN114282645A CN202111404758.8A CN202111404758A CN114282645A CN 114282645 A CN114282645 A CN 114282645A CN 202111404758 A CN202111404758 A CN 202111404758A CN 114282645 A CN114282645 A CN 114282645A
Authority
CN
China
Prior art keywords
task
worker
tasks
crowdsourcing
dqn
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111404758.8A
Other languages
Chinese (zh)
Other versions
CN114282645B (en
Inventor
彭占魁
李玉
殷昱煜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Dianzi University
Original Assignee
Hangzhou Dianzi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Dianzi University filed Critical Hangzhou Dianzi University
Priority to CN202111404758.8A priority Critical patent/CN114282645B/en
Publication of CN114282645A publication Critical patent/CN114282645A/en
Application granted granted Critical
Publication of CN114282645B publication Critical patent/CN114282645B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a DQN-based space-time crowdsourcing task allocation method. The invention aims to provide an allocation strategy for space-time crowdsourcing tasks. Extracting feature vectors according to the history records of crowdsourcing initiators and workers, and continuously training in a neural network structure designed based on DQN, so that the characteristics and the preference of the crowdsourcing initiators and the workers can be better identified, and the distribution of space-time crowdsourcing tasks is completed.

Description

DQN-based space-time crowdsourcing task allocation method
Technical Field
The invention relates to the field of space-time task crowdsourcing distribution, in particular to a method for distributing space-time crowdsourcing tasks based on DQN (deep Q network).
Background
Crowdsourcing refers to the practice of "a company or organization outsourcing work tasks performed by employees in the past to an unspecified (and often large) mass network in a free-voluntary manner; crowdsourced tasks are typically undertaken by individuals, but may also occur in a form that relies on the production of an open source individual if it involves a task that requires multiple persons to collaborate to accomplish.
Spatial crowd sourcing is the process of crowd sourcing a set of spatial tasks to a set of workers, which requires that the workers are physically located at the location to perform the corresponding tasks. The flow of the spatial crowdsourcing task is as follows: 1. a Requester (Requester) submits its own task and its information accompanying the task to a crowdsourcing platform (agent). 2. The platform publishes a set of location-related tasks in a crowd-sourced form to some crowd-sourced practitioners (Worker). 3. The crowdsourcing practitioners accept the tasks delivered by the platform and need to go to the designated locations to perform the crowdsourcing tasks. Many practical problems are therefore modeled as crowd-sourced task problems to solve.
DQN is a deep reinforcement learning, an algorithm combining deep learning and reinforcement learning, and DQN is a variant of Q-learning algorithm. Q-Learning is a value-based algorithm in a reinforcement Learning algorithm, Q is Q (S, a), namely in an S state (S belongs to S) at a certain moment, an action a (a belongs to A) is taken to obtain the expectation of profit, and the environment feeds back corresponding reward according to the action of agent. The iterative formula for Q-learning is as follows:
Q(si,ai)←(si,ai)+α[ri+1+γmaxQ(si+1ai+1,)-Q(si,ai)]
Q(si,ai) Is the state and action at time i, r is the actual gain of the current action, γ is the attenuation, maxQ(s)i+1ai+1And) is the maximum value in the Q table that can be obtained when s' is selected from the Q table, and α is the learning degree.
The DQN adds a neural network on the basis of Q-learning, and is different from other machine learning algorithms, and the DQN can solve the problems of random conversion and reward without adjustment. DQN combines the convolutional neural network with Q-learning and introduces an empirical playback mechanism, so that a computer can learn a control strategy directly according to high-dimensional perception input.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides a DQN-based space-time crowdsourcing task allocation method.
The invention comprises the following steps:
step 1, first, from the ringSituation acquisition to worker (W)i) And its selectable task list (T)i). The method specifically comprises the following substeps:
step 1.1, at the moment i, a requester issues a space-time crowdsourcing task to a crowdsourcing platform.
Step 1.2, at a certain worker WiAfter arrival, a series of crowdsourcing tasks are obtained.
Step 1.3, according to a series of simple constraints (removing too far distance, completed tasks and the like), screening to form an optional task list Ti
Step 2, extracting the worker WiFeature vector and optional task list T ofiAre connected into a feature vector fsi. The method specifically comprises the following substeps:
step 2.1, extracting optional task list TiThe feature vector of (2). Current research generally considers that the allocation strategy of spatio-temporal crowdsourcing task needs to consider factors such as return, cost, task type and current position. Therefore, we also consider the above factors. As shown in table 1: for the reward, the revenue represented as the task is ranked among all available tasks. For cost, consider two parts, one is the pick-up cost (i.e., the ranking of the acquisition task among all tasks) and one is the delivery cost (i.e., the ranking of the completion task among all tasks). For task types, it is considered that tasks are initiated at different stages of the day (such as morning, noon and evening). For the positions, they are encoded by the geohash method. Expressed as: f. ofti
Type (B) Means of
Reporting Ranking of the benefits of the task among all available tasks
Cost of Pick-up and delivery costs
Task type The task is initiated at different stages of the day
Position of Encoding them by the geohash method
TABLE 1 composition of feature vectors
Step 2.2, extracting worker WiThe feature vector of (2). For worker WiFeature vector, worker WiRecently completed task features may be used to simulate the computation of worker WiiProbability of completing the task in the future. Therefore, the worker WiExpressed as the mean of the recently completed task features, expressed as fwi. Meanwhile, in the completed task, the worker WiThe closer the time of completion is to time i, the higher the occupancy of the feature vector. The expression formula is:
Figure BDA0003372420300000021
wherein n represents a worker WiTotal number of recently completed tasks (in chronological order, from small to large), TkTask completed the kth, (1/2)n-kRepresenting the attenuation factor.
Step 2.3, putting the worker WiAnd an optional task list TiAre combined into a complete feature vector fsi. By combining ftiAnd fwiTo obtain fsi. Will f issiAs input to the DQN. At ftiThe characteristics of the tasks comprise the ranking of the rewards and costs, so the characteristic values of different available tasks are different. At fwiMiddle worker WiIs characterized by being completed byWhile taking into account the decay over time. Thus, fwiRepresents a worker WiA property when processing a task. The number of optional tasks is different at different times, so the maximum number of tasks is limited (set to maxT) and when the number of optional tasks is insufficient, padding with 0 to fix fsiThe number of the cells. Thus, fsiThe number of (2) is maxT.
Step 3, predicting recommendation W through a neural networkiThe possibility of (a). The method comprises the following specific steps:
step 3.1, feature vector fsiPut into Q network (Q-network (W)) representing worker and Q network (Q-network (R)) representing requester, respectively, and predict recommended behavior (a)i) Scores in both networks. The structure of both networks is similar, as shown in fig. 2: composed of two types of Layers, i.e. Linear Layers (row-wise) and Attention Layers (multi-head), the feature vector fsiAnd converted to a Q value. The Linear Layers is a feed-forward layer (rFF: row-wise fed-forward layer), and the Attention Layers is an Attention mechanism layer. Input feature vector fsiRepresents a worker WiAnd an optional task list TiThe characteristics of (1). The initial Linear Layers are used to apply the feature vector fsiInto higher dimensional features. The conversion formula is as follows:
rFF(X)=relu(XW+b)
x is the input, W and b are the learning parameters, relu is the activation function. The Attention Layers are used to calculate: worker WiAnd optional task list TiWith different combined feature weights. Here, two types of attachment Layers, soft-attachment layer and self-attachment layer, are used to achieve more accurate feature weights.
Then, after the above-mentioned Attention Layers, a Linear Layers is used. Helping to keep the network stable.
Next, two Attention Layers are again used, which enables the Q-Network to compute the worker WiAnd optional task list TiHigh order pairwise interactions between.
The final layer of Linear Layers reduces the characteristics of each element to a value.
And 3.2, after the scores of the two Q-networks are weighted and averaged, sorting the scores into a recommendation list.
Step 4, as worker WiAfter obtaining the recommendation list, we consider WiAnd after browsing in sequence, selecting one of the browsing to finish.
Step 5, according to the worker WiUpon completion of the task, the recommendation list is quantified as an evaluation (r)i)。
Step 6, counting the successful transfer process (S)i,ai,ri,Si+1) (when a)iBy worker WiSelection, successfully completed task) and failed transfer process (S)i,ai,0,Si+1)(SiIs represented by WiCurrent state, aiIs not processed by worker WiSelecting the completed task, Si+1Represents WiState after performing the action), it is put into a training pool (memory pool).
And 7, using the data in the training pool for Q-network (W) training.
And 8, using the data in the training pool for Q-network (R) training.
The invention has the beneficial effects that: the invention provides a DQN-based space-time crowdsourcing task allocation method, and aims to provide an allocation strategy for space-time crowdsourcing tasks. Extracting feature vectors according to historical records of crowdsourcing initiators and workers, and continuously training in a neural network structure designed based on DQN, so that the network can better identify the characteristics and preferences of the crowdsourcing initiators and the workers to complete the distribution of space-time crowdsourcing tasks.
Drawings
FIG. 1 is a block diagram of the DAIN system;
FIG. 2 is a Q-network architecture diagram of the DAIN;
FIG. 3 is a comparative CR for Workers diagram;
FIG. 4 nDCG-CR for Workers comparison chart;
FIG. 5 is a comparative graph of CR for requests;
FIG. 6 nDCG-CR for requests comparison scheme;
FIG. 7 is a comparative CR for Balance diagram;
FIG. 8 nDCG-CR for Balance comparison chart.
Detailed Description
The invention provides a DQN-based space-time crowdsourcing task allocation method, and aims to provide an allocation strategy for space-time crowdsourcing tasks. And designing a neural network structure based on the DQN, extracting feature vectors for continuous training according to the history of crowdsourcing initiators and workers, and identifying the characteristics and the preference of the crowdsourcing initiators and the workers to complete the distribution of space-time crowdsourcing tasks.
In spatiotemporal crowdsourcing tasks, principals typically have Requesters (Requesters), crowdsourcing platforms (agents), and Workers (Workers). The working process is as follows; at some point, the requestor initiates a spatio-temporal crowdsourcing task. The platform allocates proper workers for the space-time crowdsourcing task, and the workers complete the task.
For the above problem, the present invention models a Markov Decision Process (Markov Decision Process) so that the allocation strategy is optimized.
FIG. 1 is a block diagram of a DAIN (deep Adaptive Interest Network for Task Assignment in Spatial Crowndsourceing) system. At time i, the requestor issues a spatiotemporal crowdsourcing task into the crowdsourcing platform. In a worker (W)i) When the task arrives, a series of crowdsourcing tasks (T) are acquiredi). And screening the task list into an optional task list according to a series of constraints (step 1). Extraction worker WiFeature vector and optional task list T ofiAre connected into a feature vector fsi(step 2). Then, the feature vector f is divided intosiPut into two DQNs (Q-network (W) and Q-network (R) on behalf of worker and requester, respectively), each behavior a is predictediFraction of two DQN. The two scores are combined and sorted into a recommendation list (step 3). When worker WiAfter seeing the recommendation list, assume worker WiBrowsing in sequence, and selecting one of them to finish(step 4). According to the worker WiUpon completion of the task, the recommendation list is quantified as an evaluation (r)i) (step 5), for Q-network (W), riThe calculation formula is as follows:
Figure BDA0003372420300000031
Figure BDA0003372420300000032
for Q-network (R), riThe calculation formula is as follows:
Figure BDA0003372420300000033
next, the successful transfer process is counted (S)i,ai,ri,Si+1) (when a)iIs worked by worker WiSelection, successfully completed task) and failed transfer process (S)i,ai,0,Si+1) (when a)iIs not processed by worker WiThe completed task is selected) and placed in a training pool (memory pool) (step 6). The data in the training pool is used for the training of Q-network (W) and Q-network (R) (steps 7, 8).
Worker WiRecently completed task features are used to simulate the computation of worker WiProbability of completing the task in the future. Therefore, the worker WiIs represented as a weighted average of its most recently completed features.
Current research generally considers that the allocation strategy of spatio-temporal crowdsourcing task needs to consider factors such as return, cost, task type and current position. As shown in table 1: for the reward, the revenue represented as the task is ranked among all available tasks. For cost, consider two parts, one is the pick-up cost (i.e., the ranking of the acquisition task among all tasks) and one is the delivery cost (i.e., the ranking of the completion task among all tasks). For task types, it is considered that tasks are initiated at different stages of the day (such as morning, noon and evening). For positions, they are coded by the geohash method, expressed as: f. ofti
Type (B) Means of
Reporting Ranking of the benefits of the task among all available tasks
Cost of Pick-up and delivery costs
Task type The task is initiated at different stages of the day
Position of Encoding them by the geohash method
TABLE 1 composition of feature vectors
For worker WiFeature vector, worker WiRecently completed task features may be used to simulate the computation of worker WiProbability of completing the task in the future. Therefore, the worker WiExpressed as the mean of the most recently completed tasks, expressed as fwi. Meanwhile, in the completed task, the worker WiThe closer the time of completion is to time i, the higher the occupancy of the feature vector. The expression is as follows:
Figure BDA0003372420300000041
wherein n represents a worker WiTotal number of recently completed tasks (in chronological order, from small to large), TkTask completed the kth, (1/2)n-kRepresenting the attenuation factor.
By combining ftiAnd fwiTo obtain fsiA 1 is to fsiAs input to the DQN. At ftiThe characteristics of the tasks include the ranking of rewards and costs, so the characteristic values of different available tasks are not the same. At fwiMiddle worker WiIs characterized by the history of the task it has completed, while taking into account the decay over time. Thus, fwiRepresents a worker WiA property when processing a task. The number of optional tasks is different at different times, so the maximum number of tasks is limited (set to maxT) and when the number of optional tasks is insufficient, padding with 0 to fix fsiThe number of the cells. Thus, fsiThe number of (2) is maxT.
As shown in FIG. 2, Linear Layers and Attention Layers are used to apply the feature vector fsiAnd converted to a Q value. The Linear Layers is a row-wise fed-forward layer, and the Attention Layers is an Attention mechanism. Input feature vector fsiRepresents a worker WiAnd an optional task list TiThe characteristics of (1). The initial Linear Layers are used to apply the feature vector fsiTo a higher dimensional trait. The conversion formula is as follows:
rFF(X)=relu(XW+b)
x is the input, W and b are the learning parameters, relu is the activation function. Attention Layers for computing worker WiAnd optional task list TiWith different combined feature weights. Both soft-attribute layer and self-attribute layer are used to achieve more accurate feature weights. Then, adding rFF layer on the basis of original characteristics helps to keep the network stable. Then, again two Attention Layers are used, which enables the Q-Network to compute high order pairwise interactions between them, the last rFF layer reducing the characteristics of each element to a value.
The above is the main idea of the present invention, and then the validity of the method is verified on the real data set. The experimental data set used data of about 30000 tasks in west ann, china in 2016 and 10 months, provided by the dribble trip (Didi Chuxing) cover data open program. Each piece of data includes: driver ID, order ID, timestamp, longitude and latitude. From these data, information of each task is obtained. Considering that a task or a task list is recommended, the following manner is selected as the evaluation criterion:
CR (worker Completion Rate), when a worker arrives, the agent recommends a task or task list. If the task or the task list is consistent with the set task list, the task or the task list is 1;
nDCG-CR (normalized discrete temporal gain), NDCG represents the normalized loss-accumulation gain. The calculation formula is as follows:
Figure BDA0003372420300000051
the present invention will be compared with four other types of methods. All these methods are trained in real data. The methods are respectively random, Greedy + cosine similarity, neural network Greedy + neural network and DQN-based neural network, the methods calculate the return of tasks, and a proper task is selected for recommendation by predicting the return of the calculation tasks.
Random: a task recommendation is randomly selected for it in the selectable task list.
Greedy CS: the cosine similarity between the worker features and the task features is used as a completion rate, and the tasks are greedy selected or ordered according to the completion rate.
Greedy NN: the worker and task features are input into a neural network of two hidden layers to predict completion rates.
DDQN: two DQNs, based on a framework of deep reinforcement learning, maximize the benefit of the worker and requester, respectively. The Q-network is composed of a layer of Linear Layers and Attention Layers.
The model of the invention consists of two DQNs, and through experimental evaluation, the number of Q-network neurons is set to be 160. For other super-parameters, the settings are as in table 2:
radix Ginseng Value of
Number of neurons 160
target Q update frequency 50 (time) update
Size of Buffer 500
Learning rate 0.002
Reward gamma 0.98
Data size of training batch 64
TABLE 2 Superreference setup
The pytorech is used to implement the entire algorithm, and the code runs on a block of GeForce GTX 2080 Ti GPU. In terms of efficiency, the DAIN model of the present invention is similar to the comparative DDQN.
Since the data provided by the drip line does not take into account the diverse preferences of the workers, the present invention requires that it be pre-processed for evaluation. Workers are mainly divided into 4 types:
long-distance orders are favored and traffic congestion is rejected.
The short-distance order is favored, and the traffic jam is refused.
The long-distance order is favored, and the traffic jam can be accepted.
The short-distance order is liked, and the traffic jam can be accepted.
Each worker is randomly set its type of preference and sets its "selected" tasks as tasks in its set of available tasks that match the preference.
As shown in fig. 3-8, the DAIN model of the present invention is divided into two DQNs, and therefore, DQN (r) and DQN (w) are also compared with each algorithm, respectively. Random is less effective because it does not predict whether the worker will complete the task. Both Greedy CS and Greedy NN deal with only short-term rewards in front of the eye, which results in poor performance. These algorithms do not work effectively in the face of complex crowdsourcing tasks. The DDQN algorithm performs relatively well, but the DAIN model of the invention performs better due to the lack of consideration of the spatio-temporal behavior of the crowdsourcing task in the DDQN model.
In the DAIN model of the invention, the task order is sorted according to the Q values summarized by Q-network (W) and Q-network (R). In fig. 7 and 8, performance is evaluated in terms of overall benefit. Clearly, both Random and Greedy are not effective. DDQN achieves good results by modeling complex relationships between workers and tasks through neural networks, predicting current and future rewards, and updating parameters, but the model structure of DDQN is relatively simple for handling complex spatio-temporal crowdsourcing tasks. Compared with DDQN, the DQN structure of the DAIN model is more complex, and the experimental performance is better. The DAIN model not only has the advantages of the DDQN, but also is more suitable for processing space-time crowdsourcing tasks.

Claims (4)

1. A DQN-based space-time crowdsourcing task allocation method is characterized by comprising the following steps:
step 1, obtaining a worker W from an environmentiAnd its selectable task list Ti
Step 2, extracting the worker WiSum of feature vectors ofTask selection list TiAre connected into a feature vector fsiThe method comprises the following specific steps:
step 2.1, extracting optional task list TiCharacteristic vector f ofti
For the return, expressed as a ranking of the revenue for that task among all available tasks;
regarding the cost, two parts are considered, one part is the picking cost, namely the ranking of the tasks in all the tasks is obtained; another part is the delivery cost, i.e. the ranking of the completed task among all tasks;
for the task type, consider that the task initiates in different stages of a day;
for the positions, they are encoded by the geohash method;
step 2.2, extracting worker WiThe feature vector of (2);
will worker WiCharacteristic vector f ofwiA weighted average of features expressed as their most recently completed tasks;
step 2.3, putting the worker WiAnd an optional task list TiAre combined into a complete feature vector fsi(ii) a By combining ftiAnd fwiTo obtain fsi(ii) a Will f issiAs input to DQN;
step 3, predicting recommendation W through a neural networkiThe specific steps are as follows:
step 3.1, feature vector fsiPut into Q network representing worker and Q network representing requester respectively, and predict recommended behavior (a)i) Scores at both networks;
step 3.2, after the two scores are weighted and averaged, sorting the scores into a recommendation list;
step 4, as worker WiAfter obtaining the recommendation list, consider WiAfter browsing in sequence, selecting one of the browsing;
step 5, according to the worker WiUpon completion of the task, the recommendation list is quantified as an evaluation (r)i);
Step 6, counting the successful transfer process (S)i,ai,ri,Si+1) And failed transfer procedure (S)i,ai,0,Si+1) Putting the mixture into a training pool (memory pool);
step 7, training a Q network representing a worker by using the data in the training pool;
and 8, training the Q network representing the requester by using the data in the training pool.
2. The DQN-based space-time crowdsourcing task allocation method according to claim 1, wherein the step 1 specifically comprises:
step 1.1, at a moment i, a requester issues a space-time crowdsourcing task to a crowdsourcing platform;
step 1.2, at a certain worker WiAcquiring a series of crowdsourcing tasks after arrival;
step 1.3, screening the task list into an optional task list T according to the constrainti
3. The DQN-based space-time crowdsourcing task distribution method according to claim 2, characterized in that in step 2.2: in the completed task, the worker WiThe closer the time of completion is to time i, the higher the occupancy of the feature vector.
4. The DQN-based space-time crowdsourcing task distribution method according to claim 1, characterized in that in step 2.2:
at different moments, the number of the selectable tasks is different, the maximum number of the tasks is limited, and the maximum number is set to be maxT;
when the number of optional tasks is not sufficient, fill with 0 to fix fsiThe number of the cells.
CN202111404758.8A 2021-11-24 2021-11-24 Space-time crowdsourcing task allocation method based on DQN Active CN114282645B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111404758.8A CN114282645B (en) 2021-11-24 2021-11-24 Space-time crowdsourcing task allocation method based on DQN

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111404758.8A CN114282645B (en) 2021-11-24 2021-11-24 Space-time crowdsourcing task allocation method based on DQN

Publications (2)

Publication Number Publication Date
CN114282645A true CN114282645A (en) 2022-04-05
CN114282645B CN114282645B (en) 2023-04-21

Family

ID=80870016

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111404758.8A Active CN114282645B (en) 2021-11-24 2021-11-24 Space-time crowdsourcing task allocation method based on DQN

Country Status (1)

Country Link
CN (1) CN114282645B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116302448A (en) * 2023-05-12 2023-06-23 中国科学技术大学先进技术研究院 Task scheduling method and system

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108596335A (en) * 2018-04-20 2018-09-28 浙江大学 A kind of adaptive crowdsourcing method based on deeply study
CN110430547A (en) * 2019-07-24 2019-11-08 河海大学常州校区 More AUV collaboration data collection algorithms in UASNs based on Q-learning
CN110554964A (en) * 2019-09-03 2019-12-10 大连海事大学 Web service crowdsourcing test task allocation method based on deep reinforcement learning
CN110737529A (en) * 2019-09-05 2020-01-31 北京理工大学 cluster scheduling adaptive configuration method for short-time multiple variable-size data jobs
CN111311115A (en) * 2020-03-12 2020-06-19 电子科技大学 Group task allocation method based on space crowdsourcing social influence preference
CN112328914A (en) * 2020-11-06 2021-02-05 辽宁工程技术大学 Task allocation method based on space-time crowdsourcing worker behavior prediction
CN112541037A (en) * 2020-11-25 2021-03-23 福建师范大学 Spatial crowdsourcing method and terminal based on block chain and deep reinforcement learning
CN112819210A (en) * 2021-01-20 2021-05-18 杭州电子科技大学 Online single-point task allocation method capable of being rejected by workers in space crowdsourcing
CN112965499A (en) * 2021-03-08 2021-06-15 哈尔滨工业大学(深圳) Unmanned vehicle driving decision-making method based on attention model and deep reinforcement learning
US20210357731A1 (en) * 2018-11-16 2021-11-18 Deepmind Technologies Limited Controlling agents using amortized q learning

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108596335A (en) * 2018-04-20 2018-09-28 浙江大学 A kind of adaptive crowdsourcing method based on deeply study
US20210357731A1 (en) * 2018-11-16 2021-11-18 Deepmind Technologies Limited Controlling agents using amortized q learning
CN110430547A (en) * 2019-07-24 2019-11-08 河海大学常州校区 More AUV collaboration data collection algorithms in UASNs based on Q-learning
CN110554964A (en) * 2019-09-03 2019-12-10 大连海事大学 Web service crowdsourcing test task allocation method based on deep reinforcement learning
CN110737529A (en) * 2019-09-05 2020-01-31 北京理工大学 cluster scheduling adaptive configuration method for short-time multiple variable-size data jobs
CN111311115A (en) * 2020-03-12 2020-06-19 电子科技大学 Group task allocation method based on space crowdsourcing social influence preference
CN112328914A (en) * 2020-11-06 2021-02-05 辽宁工程技术大学 Task allocation method based on space-time crowdsourcing worker behavior prediction
CN112541037A (en) * 2020-11-25 2021-03-23 福建师范大学 Spatial crowdsourcing method and terminal based on block chain and deep reinforcement learning
CN112819210A (en) * 2021-01-20 2021-05-18 杭州电子科技大学 Online single-point task allocation method capable of being rejected by workers in space crowdsourcing
CN112965499A (en) * 2021-03-08 2021-06-15 哈尔滨工业大学(深圳) Unmanned vehicle driving decision-making method based on attention model and deep reinforcement learning

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
CAIHUA SHAN等: "《An End-to-End Deep RL Framework for Task Arrangement in Crowdsourcing Platforms》", 《ARXIV》 *
SNEHA CHAUDHARI等: "《Attention Mechanism In Deep Learning》", 《ARXIV》 *
林荟荟 等: "《空间众包中可拒绝情况下的在线任务分配》", 《浙江科技学院学报》 *
王元斗: "《基于Deep-Q-network的多智能体强化学习的多目标工作流调度方法研究》", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116302448A (en) * 2023-05-12 2023-06-23 中国科学技术大学先进技术研究院 Task scheduling method and system
CN116302448B (en) * 2023-05-12 2023-08-11 中国科学技术大学先进技术研究院 Task scheduling method and system

Also Published As

Publication number Publication date
CN114282645B (en) 2023-04-21

Similar Documents

Publication Publication Date Title
Zhang et al. A discounted trade reduction mechanism for dynamic ridesharing pricing
CN110400128B (en) Spatial crowdsourcing task allocation method based on worker preference perception
CN104468413B (en) A kind of network service method and system
CN110415036B (en) User grade determining method, device, computer equipment and storage medium
CN106453608A (en) Cloud-based adaptive scheduling algorithm for background request of mobile application
US20220327495A1 (en) Intelligent scheduling using a prediction model
CN108182633A (en) Loan data processing method, device, computer equipment and storage medium
CN114282645A (en) DQN-based space-time crowdsourcing task allocation method
US11727137B2 (en) Predictive update of profile database with update logic
JP2019074988A (en) Room charge setting device, room charge setting method, and program
US20140344020A1 (en) Competitor pricing strategy determination
US20040133393A1 (en) Prediction system based on weighted expert opinions using prior success measures
CN108694182A (en) Movable method for pushing, movable pusher and server
CN109872058B (en) Multimedia crowd sensing excitation method for machine learning system
CN115423393B (en) Order scheduling method and device of dynamic self-adaptive scheduling period based on LSTM
CN116595364A (en) Training method of flow prediction model, flow prediction method and device
CN115713128A (en) Federal learning method based on equipment training time fairness
US20220309438A1 (en) Automated claims adjustment assignment utilizing crowdsourcing and adjuster priority score determinations
US20140344021A1 (en) Reactive competitor price determination using a competitor response model
CN115018428A (en) Distribution scheduling method and device considering prediction uncertainty and storage medium
CN110322055A (en) A kind of method and system improving data risk model scoring stability
CN117422266B (en) Task allocation method, system, device and storage medium based on worker preference
CN111612419A (en) Method and device for processing power declaration data and computer equipment
CN113298629B (en) Data sharing method based on two-way auction theory and intelligent contract
US11425250B1 (en) Artificial intelligence based call handling optimization

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB03 Change of inventor or designer information

Inventor after: Yin Yuyu

Inventor after: Li Yu

Inventor after: Peng Zhankui

Inventor before: Peng Zhankui

Inventor before: Li Yu

Inventor before: Yin Yuyu

CB03 Change of inventor or designer information
GR01 Patent grant
GR01 Patent grant