CN114282645A - DQN-based space-time crowdsourcing task allocation method - Google Patents
DQN-based space-time crowdsourcing task allocation method Download PDFInfo
- Publication number
- CN114282645A CN114282645A CN202111404758.8A CN202111404758A CN114282645A CN 114282645 A CN114282645 A CN 114282645A CN 202111404758 A CN202111404758 A CN 202111404758A CN 114282645 A CN114282645 A CN 114282645A
- Authority
- CN
- China
- Prior art keywords
- task
- worker
- tasks
- crowdsourcing
- dqn
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
The invention discloses a DQN-based space-time crowdsourcing task allocation method. The invention aims to provide an allocation strategy for space-time crowdsourcing tasks. Extracting feature vectors according to the history records of crowdsourcing initiators and workers, and continuously training in a neural network structure designed based on DQN, so that the characteristics and the preference of the crowdsourcing initiators and the workers can be better identified, and the distribution of space-time crowdsourcing tasks is completed.
Description
Technical Field
The invention relates to the field of space-time task crowdsourcing distribution, in particular to a method for distributing space-time crowdsourcing tasks based on DQN (deep Q network).
Background
Crowdsourcing refers to the practice of "a company or organization outsourcing work tasks performed by employees in the past to an unspecified (and often large) mass network in a free-voluntary manner; crowdsourced tasks are typically undertaken by individuals, but may also occur in a form that relies on the production of an open source individual if it involves a task that requires multiple persons to collaborate to accomplish.
Spatial crowd sourcing is the process of crowd sourcing a set of spatial tasks to a set of workers, which requires that the workers are physically located at the location to perform the corresponding tasks. The flow of the spatial crowdsourcing task is as follows: 1. a Requester (Requester) submits its own task and its information accompanying the task to a crowdsourcing platform (agent). 2. The platform publishes a set of location-related tasks in a crowd-sourced form to some crowd-sourced practitioners (Worker). 3. The crowdsourcing practitioners accept the tasks delivered by the platform and need to go to the designated locations to perform the crowdsourcing tasks. Many practical problems are therefore modeled as crowd-sourced task problems to solve.
DQN is a deep reinforcement learning, an algorithm combining deep learning and reinforcement learning, and DQN is a variant of Q-learning algorithm. Q-Learning is a value-based algorithm in a reinforcement Learning algorithm, Q is Q (S, a), namely in an S state (S belongs to S) at a certain moment, an action a (a belongs to A) is taken to obtain the expectation of profit, and the environment feeds back corresponding reward according to the action of agent. The iterative formula for Q-learning is as follows:
Q(si,ai)←(si,ai)+α[ri+1+γmaxQ(si+1ai+1,)-Q(si,ai)]
Q(si,ai) Is the state and action at time i, r is the actual gain of the current action, γ is the attenuation, maxQ(s)i+1ai+1And) is the maximum value in the Q table that can be obtained when s' is selected from the Q table, and α is the learning degree.
The DQN adds a neural network on the basis of Q-learning, and is different from other machine learning algorithms, and the DQN can solve the problems of random conversion and reward without adjustment. DQN combines the convolutional neural network with Q-learning and introduces an empirical playback mechanism, so that a computer can learn a control strategy directly according to high-dimensional perception input.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides a DQN-based space-time crowdsourcing task allocation method.
The invention comprises the following steps:
step 1.1, at the moment i, a requester issues a space-time crowdsourcing task to a crowdsourcing platform.
Step 1.2, at a certain worker WiAfter arrival, a series of crowdsourcing tasks are obtained.
Step 1.3, according to a series of simple constraints (removing too far distance, completed tasks and the like), screening to form an optional task list Ti。
step 2.1, extracting optional task list TiThe feature vector of (2). Current research generally considers that the allocation strategy of spatio-temporal crowdsourcing task needs to consider factors such as return, cost, task type and current position. Therefore, we also consider the above factors. As shown in table 1: for the reward, the revenue represented as the task is ranked among all available tasks. For cost, consider two parts, one is the pick-up cost (i.e., the ranking of the acquisition task among all tasks) and one is the delivery cost (i.e., the ranking of the completion task among all tasks). For task types, it is considered that tasks are initiated at different stages of the day (such as morning, noon and evening). For the positions, they are encoded by the geohash method. Expressed as: f. ofti。
Type (B) | Means of |
Reporting | Ranking of the benefits of the task among all available tasks |
Cost of | Pick-up and delivery costs |
Task type | The task is initiated at different stages of the day |
Position of | Encoding them by the geohash method |
TABLE 1 composition of feature vectors
Step 2.2, extracting worker WiThe feature vector of (2). For worker WiFeature vector, worker WiRecently completed task features may be used to simulate the computation of worker WiiProbability of completing the task in the future. Therefore, the worker WiExpressed as the mean of the recently completed task features, expressed as fwi. Meanwhile, in the completed task, the worker WiThe closer the time of completion is to time i, the higher the occupancy of the feature vector. The expression formula is:wherein n represents a worker WiTotal number of recently completed tasks (in chronological order, from small to large), TkTask completed the kth, (1/2)n-kRepresenting the attenuation factor.
Step 2.3, putting the worker WiAnd an optional task list TiAre combined into a complete feature vector fsi. By combining ftiAnd fwiTo obtain fsi. Will f issiAs input to the DQN. At ftiThe characteristics of the tasks comprise the ranking of the rewards and costs, so the characteristic values of different available tasks are different. At fwiMiddle worker WiIs characterized by being completed byWhile taking into account the decay over time. Thus, fwiRepresents a worker WiA property when processing a task. The number of optional tasks is different at different times, so the maximum number of tasks is limited (set to maxT) and when the number of optional tasks is insufficient, padding with 0 to fix fsiThe number of the cells. Thus, fsiThe number of (2) is maxT.
Step 3, predicting recommendation W through a neural networkiThe possibility of (a). The method comprises the following specific steps:
step 3.1, feature vector fsiPut into Q network (Q-network (W)) representing worker and Q network (Q-network (R)) representing requester, respectively, and predict recommended behavior (a)i) Scores in both networks. The structure of both networks is similar, as shown in fig. 2: composed of two types of Layers, i.e. Linear Layers (row-wise) and Attention Layers (multi-head), the feature vector fsiAnd converted to a Q value. The Linear Layers is a feed-forward layer (rFF: row-wise fed-forward layer), and the Attention Layers is an Attention mechanism layer. Input feature vector fsiRepresents a worker WiAnd an optional task list TiThe characteristics of (1). The initial Linear Layers are used to apply the feature vector fsiInto higher dimensional features. The conversion formula is as follows:
rFF(X)=relu(XW+b)
x is the input, W and b are the learning parameters, relu is the activation function. The Attention Layers are used to calculate: worker WiAnd optional task list TiWith different combined feature weights. Here, two types of attachment Layers, soft-attachment layer and self-attachment layer, are used to achieve more accurate feature weights.
Then, after the above-mentioned Attention Layers, a Linear Layers is used. Helping to keep the network stable.
Next, two Attention Layers are again used, which enables the Q-Network to compute the worker WiAnd optional task list TiHigh order pairwise interactions between.
The final layer of Linear Layers reduces the characteristics of each element to a value.
And 3.2, after the scores of the two Q-networks are weighted and averaged, sorting the scores into a recommendation list.
And 7, using the data in the training pool for Q-network (W) training.
And 8, using the data in the training pool for Q-network (R) training.
The invention has the beneficial effects that: the invention provides a DQN-based space-time crowdsourcing task allocation method, and aims to provide an allocation strategy for space-time crowdsourcing tasks. Extracting feature vectors according to historical records of crowdsourcing initiators and workers, and continuously training in a neural network structure designed based on DQN, so that the network can better identify the characteristics and preferences of the crowdsourcing initiators and the workers to complete the distribution of space-time crowdsourcing tasks.
Drawings
FIG. 1 is a block diagram of the DAIN system;
FIG. 2 is a Q-network architecture diagram of the DAIN;
FIG. 3 is a comparative CR for Workers diagram;
FIG. 4 nDCG-CR for Workers comparison chart;
FIG. 5 is a comparative graph of CR for requests;
FIG. 6 nDCG-CR for requests comparison scheme;
FIG. 7 is a comparative CR for Balance diagram;
FIG. 8 nDCG-CR for Balance comparison chart.
Detailed Description
The invention provides a DQN-based space-time crowdsourcing task allocation method, and aims to provide an allocation strategy for space-time crowdsourcing tasks. And designing a neural network structure based on the DQN, extracting feature vectors for continuous training according to the history of crowdsourcing initiators and workers, and identifying the characteristics and the preference of the crowdsourcing initiators and the workers to complete the distribution of space-time crowdsourcing tasks.
In spatiotemporal crowdsourcing tasks, principals typically have Requesters (Requesters), crowdsourcing platforms (agents), and Workers (Workers). The working process is as follows; at some point, the requestor initiates a spatio-temporal crowdsourcing task. The platform allocates proper workers for the space-time crowdsourcing task, and the workers complete the task.
For the above problem, the present invention models a Markov Decision Process (Markov Decision Process) so that the allocation strategy is optimized.
FIG. 1 is a block diagram of a DAIN (deep Adaptive Interest Network for Task Assignment in Spatial Crowndsourceing) system. At time i, the requestor issues a spatiotemporal crowdsourcing task into the crowdsourcing platform. In a worker (W)i) When the task arrives, a series of crowdsourcing tasks (T) are acquiredi). And screening the task list into an optional task list according to a series of constraints (step 1). Extraction worker WiFeature vector and optional task list T ofiAre connected into a feature vector fsi(step 2). Then, the feature vector f is divided intosiPut into two DQNs (Q-network (W) and Q-network (R) on behalf of worker and requester, respectively), each behavior a is predictediFraction of two DQN. The two scores are combined and sorted into a recommendation list (step 3). When worker WiAfter seeing the recommendation list, assume worker WiBrowsing in sequence, and selecting one of them to finish(step 4). According to the worker WiUpon completion of the task, the recommendation list is quantified as an evaluation (r)i) (step 5), for Q-network (W), riThe calculation formula is as follows: for Q-network (R), riThe calculation formula is as follows:next, the successful transfer process is counted (S)i,ai,ri,Si+1) (when a)iIs worked by worker WiSelection, successfully completed task) and failed transfer process (S)i,ai,0,Si+1) (when a)iIs not processed by worker WiThe completed task is selected) and placed in a training pool (memory pool) (step 6). The data in the training pool is used for the training of Q-network (W) and Q-network (R) (steps 7, 8).
Worker WiRecently completed task features are used to simulate the computation of worker WiProbability of completing the task in the future. Therefore, the worker WiIs represented as a weighted average of its most recently completed features.
Current research generally considers that the allocation strategy of spatio-temporal crowdsourcing task needs to consider factors such as return, cost, task type and current position. As shown in table 1: for the reward, the revenue represented as the task is ranked among all available tasks. For cost, consider two parts, one is the pick-up cost (i.e., the ranking of the acquisition task among all tasks) and one is the delivery cost (i.e., the ranking of the completion task among all tasks). For task types, it is considered that tasks are initiated at different stages of the day (such as morning, noon and evening). For positions, they are coded by the geohash method, expressed as: f. ofti。
Type (B) | Means of |
Reporting | Ranking of the benefits of the task among all available tasks |
Cost of | Pick-up and delivery costs |
Task type | The task is initiated at different stages of the day |
Position of | Encoding them by the geohash method |
TABLE 1 composition of feature vectors
For worker WiFeature vector, worker WiRecently completed task features may be used to simulate the computation of worker WiProbability of completing the task in the future. Therefore, the worker WiExpressed as the mean of the most recently completed tasks, expressed as fwi. Meanwhile, in the completed task, the worker WiThe closer the time of completion is to time i, the higher the occupancy of the feature vector. The expression is as follows:
wherein n represents a worker WiTotal number of recently completed tasks (in chronological order, from small to large), TkTask completed the kth, (1/2)n-kRepresenting the attenuation factor.
By combining ftiAnd fwiTo obtain fsiA 1 is to fsiAs input to the DQN. At ftiThe characteristics of the tasks include the ranking of rewards and costs, so the characteristic values of different available tasks are not the same. At fwiMiddle worker WiIs characterized by the history of the task it has completed, while taking into account the decay over time. Thus, fwiRepresents a worker WiA property when processing a task. The number of optional tasks is different at different times, so the maximum number of tasks is limited (set to maxT) and when the number of optional tasks is insufficient, padding with 0 to fix fsiThe number of the cells. Thus, fsiThe number of (2) is maxT.
As shown in FIG. 2, Linear Layers and Attention Layers are used to apply the feature vector fsiAnd converted to a Q value. The Linear Layers is a row-wise fed-forward layer, and the Attention Layers is an Attention mechanism. Input feature vector fsiRepresents a worker WiAnd an optional task list TiThe characteristics of (1). The initial Linear Layers are used to apply the feature vector fsiTo a higher dimensional trait. The conversion formula is as follows:
rFF(X)=relu(XW+b)
x is the input, W and b are the learning parameters, relu is the activation function. Attention Layers for computing worker WiAnd optional task list TiWith different combined feature weights. Both soft-attribute layer and self-attribute layer are used to achieve more accurate feature weights. Then, adding rFF layer on the basis of original characteristics helps to keep the network stable. Then, again two Attention Layers are used, which enables the Q-Network to compute high order pairwise interactions between them, the last rFF layer reducing the characteristics of each element to a value.
The above is the main idea of the present invention, and then the validity of the method is verified on the real data set. The experimental data set used data of about 30000 tasks in west ann, china in 2016 and 10 months, provided by the dribble trip (Didi Chuxing) cover data open program. Each piece of data includes: driver ID, order ID, timestamp, longitude and latitude. From these data, information of each task is obtained. Considering that a task or a task list is recommended, the following manner is selected as the evaluation criterion:
CR (worker Completion Rate), when a worker arrives, the agent recommends a task or task list. If the task or the task list is consistent with the set task list, the task or the task list is 1;
nDCG-CR (normalized discrete temporal gain), NDCG represents the normalized loss-accumulation gain. The calculation formula is as follows:
the present invention will be compared with four other types of methods. All these methods are trained in real data. The methods are respectively random, Greedy + cosine similarity, neural network Greedy + neural network and DQN-based neural network, the methods calculate the return of tasks, and a proper task is selected for recommendation by predicting the return of the calculation tasks.
Random: a task recommendation is randomly selected for it in the selectable task list.
Greedy CS: the cosine similarity between the worker features and the task features is used as a completion rate, and the tasks are greedy selected or ordered according to the completion rate.
Greedy NN: the worker and task features are input into a neural network of two hidden layers to predict completion rates.
DDQN: two DQNs, based on a framework of deep reinforcement learning, maximize the benefit of the worker and requester, respectively. The Q-network is composed of a layer of Linear Layers and Attention Layers.
The model of the invention consists of two DQNs, and through experimental evaluation, the number of Q-network neurons is set to be 160. For other super-parameters, the settings are as in table 2:
radix Ginseng | Value of |
Number of neurons | 160 |
target Q update frequency | 50 (time) update |
Size of Buffer | 500 |
Learning rate | 0.002 |
Reward gamma | 0.98 |
Data size of training batch | 64 |
TABLE 2 Superreference setup
The pytorech is used to implement the entire algorithm, and the code runs on a block of GeForce GTX 2080 Ti GPU. In terms of efficiency, the DAIN model of the present invention is similar to the comparative DDQN.
Since the data provided by the drip line does not take into account the diverse preferences of the workers, the present invention requires that it be pre-processed for evaluation. Workers are mainly divided into 4 types:
long-distance orders are favored and traffic congestion is rejected.
The short-distance order is favored, and the traffic jam is refused.
The long-distance order is favored, and the traffic jam can be accepted.
The short-distance order is liked, and the traffic jam can be accepted.
Each worker is randomly set its type of preference and sets its "selected" tasks as tasks in its set of available tasks that match the preference.
As shown in fig. 3-8, the DAIN model of the present invention is divided into two DQNs, and therefore, DQN (r) and DQN (w) are also compared with each algorithm, respectively. Random is less effective because it does not predict whether the worker will complete the task. Both Greedy CS and Greedy NN deal with only short-term rewards in front of the eye, which results in poor performance. These algorithms do not work effectively in the face of complex crowdsourcing tasks. The DDQN algorithm performs relatively well, but the DAIN model of the invention performs better due to the lack of consideration of the spatio-temporal behavior of the crowdsourcing task in the DDQN model.
In the DAIN model of the invention, the task order is sorted according to the Q values summarized by Q-network (W) and Q-network (R). In fig. 7 and 8, performance is evaluated in terms of overall benefit. Clearly, both Random and Greedy are not effective. DDQN achieves good results by modeling complex relationships between workers and tasks through neural networks, predicting current and future rewards, and updating parameters, but the model structure of DDQN is relatively simple for handling complex spatio-temporal crowdsourcing tasks. Compared with DDQN, the DQN structure of the DAIN model is more complex, and the experimental performance is better. The DAIN model not only has the advantages of the DDQN, but also is more suitable for processing space-time crowdsourcing tasks.
Claims (4)
1. A DQN-based space-time crowdsourcing task allocation method is characterized by comprising the following steps:
step 1, obtaining a worker W from an environmentiAnd its selectable task list Ti;
Step 2, extracting the worker WiSum of feature vectors ofTask selection list TiAre connected into a feature vector fsiThe method comprises the following specific steps:
step 2.1, extracting optional task list TiCharacteristic vector f ofti;
For the return, expressed as a ranking of the revenue for that task among all available tasks;
regarding the cost, two parts are considered, one part is the picking cost, namely the ranking of the tasks in all the tasks is obtained; another part is the delivery cost, i.e. the ranking of the completed task among all tasks;
for the task type, consider that the task initiates in different stages of a day;
for the positions, they are encoded by the geohash method;
step 2.2, extracting worker WiThe feature vector of (2);
will worker WiCharacteristic vector f ofwiA weighted average of features expressed as their most recently completed tasks;
step 2.3, putting the worker WiAnd an optional task list TiAre combined into a complete feature vector fsi(ii) a By combining ftiAnd fwiTo obtain fsi(ii) a Will f issiAs input to DQN;
step 3, predicting recommendation W through a neural networkiThe specific steps are as follows:
step 3.1, feature vector fsiPut into Q network representing worker and Q network representing requester respectively, and predict recommended behavior (a)i) Scores at both networks;
step 3.2, after the two scores are weighted and averaged, sorting the scores into a recommendation list;
step 4, as worker WiAfter obtaining the recommendation list, consider WiAfter browsing in sequence, selecting one of the browsing;
step 5, according to the worker WiUpon completion of the task, the recommendation list is quantified as an evaluation (r)i);
Step 6, counting the successful transfer process (S)i,ai,ri,Si+1) And failed transfer procedure (S)i,ai,0,Si+1) Putting the mixture into a training pool (memory pool);
step 7, training a Q network representing a worker by using the data in the training pool;
and 8, training the Q network representing the requester by using the data in the training pool.
2. The DQN-based space-time crowdsourcing task allocation method according to claim 1, wherein the step 1 specifically comprises:
step 1.1, at a moment i, a requester issues a space-time crowdsourcing task to a crowdsourcing platform;
step 1.2, at a certain worker WiAcquiring a series of crowdsourcing tasks after arrival;
step 1.3, screening the task list into an optional task list T according to the constrainti。
3. The DQN-based space-time crowdsourcing task distribution method according to claim 2, characterized in that in step 2.2: in the completed task, the worker WiThe closer the time of completion is to time i, the higher the occupancy of the feature vector.
4. The DQN-based space-time crowdsourcing task distribution method according to claim 1, characterized in that in step 2.2:
at different moments, the number of the selectable tasks is different, the maximum number of the tasks is limited, and the maximum number is set to be maxT;
when the number of optional tasks is not sufficient, fill with 0 to fix fsiThe number of the cells.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111404758.8A CN114282645B (en) | 2021-11-24 | 2021-11-24 | Space-time crowdsourcing task allocation method based on DQN |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111404758.8A CN114282645B (en) | 2021-11-24 | 2021-11-24 | Space-time crowdsourcing task allocation method based on DQN |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114282645A true CN114282645A (en) | 2022-04-05 |
CN114282645B CN114282645B (en) | 2023-04-21 |
Family
ID=80870016
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111404758.8A Active CN114282645B (en) | 2021-11-24 | 2021-11-24 | Space-time crowdsourcing task allocation method based on DQN |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114282645B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116302448A (en) * | 2023-05-12 | 2023-06-23 | 中国科学技术大学先进技术研究院 | Task scheduling method and system |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108596335A (en) * | 2018-04-20 | 2018-09-28 | 浙江大学 | A kind of adaptive crowdsourcing method based on deeply study |
CN110430547A (en) * | 2019-07-24 | 2019-11-08 | 河海大学常州校区 | More AUV collaboration data collection algorithms in UASNs based on Q-learning |
CN110554964A (en) * | 2019-09-03 | 2019-12-10 | 大连海事大学 | Web service crowdsourcing test task allocation method based on deep reinforcement learning |
CN110737529A (en) * | 2019-09-05 | 2020-01-31 | 北京理工大学 | cluster scheduling adaptive configuration method for short-time multiple variable-size data jobs |
CN111311115A (en) * | 2020-03-12 | 2020-06-19 | 电子科技大学 | Group task allocation method based on space crowdsourcing social influence preference |
CN112328914A (en) * | 2020-11-06 | 2021-02-05 | 辽宁工程技术大学 | Task allocation method based on space-time crowdsourcing worker behavior prediction |
CN112541037A (en) * | 2020-11-25 | 2021-03-23 | 福建师范大学 | Spatial crowdsourcing method and terminal based on block chain and deep reinforcement learning |
CN112819210A (en) * | 2021-01-20 | 2021-05-18 | 杭州电子科技大学 | Online single-point task allocation method capable of being rejected by workers in space crowdsourcing |
CN112965499A (en) * | 2021-03-08 | 2021-06-15 | 哈尔滨工业大学(深圳) | Unmanned vehicle driving decision-making method based on attention model and deep reinforcement learning |
US20210357731A1 (en) * | 2018-11-16 | 2021-11-18 | Deepmind Technologies Limited | Controlling agents using amortized q learning |
-
2021
- 2021-11-24 CN CN202111404758.8A patent/CN114282645B/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108596335A (en) * | 2018-04-20 | 2018-09-28 | 浙江大学 | A kind of adaptive crowdsourcing method based on deeply study |
US20210357731A1 (en) * | 2018-11-16 | 2021-11-18 | Deepmind Technologies Limited | Controlling agents using amortized q learning |
CN110430547A (en) * | 2019-07-24 | 2019-11-08 | 河海大学常州校区 | More AUV collaboration data collection algorithms in UASNs based on Q-learning |
CN110554964A (en) * | 2019-09-03 | 2019-12-10 | 大连海事大学 | Web service crowdsourcing test task allocation method based on deep reinforcement learning |
CN110737529A (en) * | 2019-09-05 | 2020-01-31 | 北京理工大学 | cluster scheduling adaptive configuration method for short-time multiple variable-size data jobs |
CN111311115A (en) * | 2020-03-12 | 2020-06-19 | 电子科技大学 | Group task allocation method based on space crowdsourcing social influence preference |
CN112328914A (en) * | 2020-11-06 | 2021-02-05 | 辽宁工程技术大学 | Task allocation method based on space-time crowdsourcing worker behavior prediction |
CN112541037A (en) * | 2020-11-25 | 2021-03-23 | 福建师范大学 | Spatial crowdsourcing method and terminal based on block chain and deep reinforcement learning |
CN112819210A (en) * | 2021-01-20 | 2021-05-18 | 杭州电子科技大学 | Online single-point task allocation method capable of being rejected by workers in space crowdsourcing |
CN112965499A (en) * | 2021-03-08 | 2021-06-15 | 哈尔滨工业大学(深圳) | Unmanned vehicle driving decision-making method based on attention model and deep reinforcement learning |
Non-Patent Citations (4)
Title |
---|
CAIHUA SHAN等: "《An End-to-End Deep RL Framework for Task Arrangement in Crowdsourcing Platforms》", 《ARXIV》 * |
SNEHA CHAUDHARI等: "《Attention Mechanism In Deep Learning》", 《ARXIV》 * |
林荟荟 等: "《空间众包中可拒绝情况下的在线任务分配》", 《浙江科技学院学报》 * |
王元斗: "《基于Deep-Q-network的多智能体强化学习的多目标工作流调度方法研究》", 《中国优秀硕士学位论文全文数据库 信息科技辑》 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116302448A (en) * | 2023-05-12 | 2023-06-23 | 中国科学技术大学先进技术研究院 | Task scheduling method and system |
CN116302448B (en) * | 2023-05-12 | 2023-08-11 | 中国科学技术大学先进技术研究院 | Task scheduling method and system |
Also Published As
Publication number | Publication date |
---|---|
CN114282645B (en) | 2023-04-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Zhang et al. | A discounted trade reduction mechanism for dynamic ridesharing pricing | |
CN110400128B (en) | Spatial crowdsourcing task allocation method based on worker preference perception | |
CN104468413B (en) | A kind of network service method and system | |
CN110415036B (en) | User grade determining method, device, computer equipment and storage medium | |
CN106453608A (en) | Cloud-based adaptive scheduling algorithm for background request of mobile application | |
US20220327495A1 (en) | Intelligent scheduling using a prediction model | |
CN108182633A (en) | Loan data processing method, device, computer equipment and storage medium | |
CN114282645A (en) | DQN-based space-time crowdsourcing task allocation method | |
US11727137B2 (en) | Predictive update of profile database with update logic | |
JP2019074988A (en) | Room charge setting device, room charge setting method, and program | |
US20140344020A1 (en) | Competitor pricing strategy determination | |
US20040133393A1 (en) | Prediction system based on weighted expert opinions using prior success measures | |
CN108694182A (en) | Movable method for pushing, movable pusher and server | |
CN109872058B (en) | Multimedia crowd sensing excitation method for machine learning system | |
CN115423393B (en) | Order scheduling method and device of dynamic self-adaptive scheduling period based on LSTM | |
CN116595364A (en) | Training method of flow prediction model, flow prediction method and device | |
CN115713128A (en) | Federal learning method based on equipment training time fairness | |
US20220309438A1 (en) | Automated claims adjustment assignment utilizing crowdsourcing and adjuster priority score determinations | |
US20140344021A1 (en) | Reactive competitor price determination using a competitor response model | |
CN115018428A (en) | Distribution scheduling method and device considering prediction uncertainty and storage medium | |
CN110322055A (en) | A kind of method and system improving data risk model scoring stability | |
CN117422266B (en) | Task allocation method, system, device and storage medium based on worker preference | |
CN111612419A (en) | Method and device for processing power declaration data and computer equipment | |
CN113298629B (en) | Data sharing method based on two-way auction theory and intelligent contract | |
US11425250B1 (en) | Artificial intelligence based call handling optimization |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
CB03 | Change of inventor or designer information |
Inventor after: Yin Yuyu Inventor after: Li Yu Inventor after: Peng Zhankui Inventor before: Peng Zhankui Inventor before: Li Yu Inventor before: Yin Yuyu |
|
CB03 | Change of inventor or designer information | ||
GR01 | Patent grant | ||
GR01 | Patent grant |