CN107274029A - A kind of future anticipation method of interaction medium in utilization dynamic scene - Google Patents

A kind of future anticipation method of interaction medium in utilization dynamic scene Download PDF

Info

Publication number
CN107274029A
CN107274029A CN201710487508.2A CN201710487508A CN107274029A CN 107274029 A CN107274029 A CN 107274029A CN 201710487508 A CN201710487508 A CN 201710487508A CN 107274029 A CN107274029 A CN 107274029A
Authority
CN
China
Prior art keywords
mrow
mover
medium
msubsup
msub
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN201710487508.2A
Other languages
Chinese (zh)
Inventor
夏春秋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Vision Technology Co Ltd
Original Assignee
Shenzhen Vision Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Vision Technology Co Ltd filed Critical Shenzhen Vision Technology Co Ltd
Priority to CN201710487508.2A priority Critical patent/CN107274029A/en
Publication of CN107274029A publication Critical patent/CN107274029A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks

Abstract

The future anticipation method of interaction medium in a kind of utilization dynamic scene proposed in the present invention, its main contents include:Multisample generation, sample sequence are merged with perfect, scene, its process is, the multisample for assuming prediction is generated first by condition variation self-encoding encoder, the marking and queuing of sample is carried out followed by the long-term favorable decisions method based on reverse optimum control framework and improves processing, introduce Recognition with Recurrent Neural Network and scene fusion is carried out to the medium of image-context, activity, obtained result is fed back in decoder again to be iterated.The present invention can handle the future event foresight activity of a certain medium in scene dynamically or statically and generate, sorts and improve sample there is provided a Recognition with Recurrent Neural Network set, significantly obtain the precise effect of prediction.

Description

A kind of future anticipation method of interaction medium in utilization dynamic scene
Technical field
The present invention relates to action prediction field, more particularly, to interaction medium in a kind of utilization dynamic scene not Carry out Forecasting Methodology.
Background technology
The response to predict the following scene to this track is carried out according to passing action trail, i.e., future event is predicted, is Attract the subject under discussion greatly paid close attention in recent years.Under the background of big data increased popularity, man-machine interaction, robot navigation etc. can meet Many unprecedented scenes, can high degree training smart equipment progress current scene and task with future anticipation method Real-time distribution with implement, simultaneously for the huge field of task amount, as population surveillance, resource exploration etc. possess hardware facility Scene will have high application potential, in addition, meeting an urgent need and arranging in multitask command system distributed coordination, a variety of vehicles anticollisions Apply, also have immeasurable humane and market value in terms of natural calamity escape route forewarning management.
Future anticipation is still a very highly difficult challenge.Because the scene being related in task both includes vision On dimension, considering in logic is also contains, while also assigning the function of this task Intelligent Recognition, it is necessary to high data volume Training test process.On the one hand, space-time instability causes Fuzzy Influence to visual reception, while the responsible journey of data volume Amount of calculation increase can cause the reduction of live effect caused by degree, on the other hand, and the medium interacted too much produces mutual shadow Ring, it is impossible to peel off carry out quantitative analysis one by one, therefore the prediction of action is judged to cause difficulty.
The present invention proposes a kind of long-term favorable decisions new frame based on reverse optimum control.First by condition variation Self-encoding encoder assumes the multisample of prediction to generate, followed by the long-term favorable decisions method based on reverse optimum control framework Carry out the marking and queuing of sample and improve processing, introducing Recognition with Recurrent Neural Network carries out scene to the medium of image-context, activity and melted Close, obtained result is fed back in decoder again to be iterated.The present invention can handle a certain in scene dynamically or statically The future event foresight activity of medium generates, sorts and improved sample there is provided Recognition with Recurrent Neural Network set, significantly takes The precise effect that must be predicted.
The content of the invention
The problem of for solving to predict future-action in complex scene multi interaction medium, it is an object of the invention to There is provided a kind of future anticipation method of interaction medium in utilization dynamic scene, it is proposed that a kind of based on reverse optimum control Long-term favorable decisions new frame.
To solve the above problems, the present invention provides a kind of future anticipation side of interaction medium in utilization dynamic scene Method, its main contents include:
(1) multisample is generated;
(2) sample sequence with it is perfect;
(3) scene is merged.
Wherein, described multisample generation, including universal model construction and condition variation self-encoding encoder.
Further, described universal model construction, for playing interactive object in generation event in certain scene, No matter static or motion, referred to as medium, if giving n medium, their passing movement track X=<X1,X2,…,Xn >, calculate their future behaviour track Y=<Y1,Y2,…,Yn>Respective probability then be P<Y|X,I>, especially, I is should The input at scene lower a moment, the maximum probability obtained according to input is exported as prediction locus.
Further, described condition variation self-encoding encoder, learns a decision function f using depth generation model, Passing movement track X and input I are mapped on the Y of future behaviour track by the function, specifically:
(1) random latent variable z is introducedi, in input XiUnder conditions of, study to output YiProbability distribution be P (Yi|Xi), It is respectively identification network Q that this process, which builds corresponding neutral net,φ(zi|Yi,Xi), condition pro-active network Pv(zi|Xi) and it is raw Into network Pθ(Yi|Xi,zi);
(2) in the training stage:1. respectively with two Recognition with Recurrent Neural Network to medium i, XiAnd YiEncoded, obtain two As a result 2. the two results are merged and is delivered to one layer of fully-connected network using nonlinear activation function again;3. again Generated with two parallel fully-connected networks on ziAverageAnd standard deviationGaussian distribution model is modeled as according to it, and And be normalized with KL divergences;4. propose that two loss functions are adjusted to self-encoding encoder, respectively rebuild loss function With KL divergence loss functions lKLD=DKL(Qφ(zi|Yi, Xi)||Pv(zi));
(3) in test phase:1. due to future anticipation trackDo not apply to and be rejected, therefore utilize passing track With multiple latent variable zi (k)Sample merges;2. it is relative with the training stage, nowWith β (zi (k)) it is input to Recognition with Recurrent Neural Network solution The hypothesis set of multiple predictions is produced in code device.
Further, described sample sequence with it is perfect, including sample evaluation improves and iterative feedback two parts.
Further, described sample evaluates perfect, is predicted for future event, using the decision-making in enhancing learning framework Formulating method, a medium can have selected after training one reward for a long time it is maximized action go implement as future thing Part, now designs a Recognition with Recurrent Neural Network model and goes prediction to being made in each trainingCarry out long term accumulation reward Measurement, be specially:
(1) medium i is scored, there is K forecast sampleThen the evaluation score s of each sample can Have:
Wherein,Refer to the forecast sample of other All Medias in addition to i,It is pre- test samples of the medium i in time t This,It is forecast samples all before knowing timestamp t, ψ is allocated to each time step t reward function, should Function is realized with the fully connected network network layers for being connected to Recognition with Recurrent Neural Network;
(2) perfect to medium i, in scoring process, Recognition with Recurrent Neural Network model estimates a regression vector simultaneouslyThe regression function η that the estimation procedure is used is:
Wherein, regression function η can accumulate all passing scene environments and the dynamic of all interaction mediums, while Estimated on whole time dimension optimal
Further, described iterative feedback, optimum regression substituting vector is obtained using formula (2)Go gradually complete The hypothesis set of kind predictionSpecially:
(1) each iteration,All it is updated toIt is then delivered to reverse optimum control module;
(2) reverse optimum control training in sequence and improve module, they have two kinds loss items, be respectively:Intersect Entropy loss lCE=H (p, q), wherein q are obtained by activation primitive, i.e.,Return Lose item
(3) to a training batch in neutral net, its whole multitask loss item is:
Wherein, N is the number of the batch medium.
Further, described scene fusion, including context insertion and interaction feature.
Further, described context insertion, using Recognition with Recurrent Neural Network, receives following on each timestamp t Input Xt
Wherein,It isIn the speed of t, γ is one layer of full articulamentum with nonlinear activation function, its handle Speed is mapped to higher-dimension and characterizes space,It will be located atConvolutional neural networks feature ρ (I) carry out pondization operate, R is gathered together the hiding vector on other medium spatial domains by a fused layer, self-embedding vectorIt is used as circulation nerve The initialization hidden state vector of network.
Further, described interaction feature, using the spatial domain grid operations using pond layer operation, makes different medium not With the unlikely disappearance of the interaction between sample, it is specially:In moment t medium i, its sample k spatial domain grid cell with Centered on, then to each grid cell g, the feature of its interior Hide AllAverage pond will all be carried out Change operation.
Brief description of the drawings
Fig. 1 is the system flow chart of the future anticipation method of interaction medium in a kind of utilization dynamic scene of the invention.
Fig. 2 is the example of the future anticipation method of interaction medium in a kind of utilization dynamic scene of the invention.
Embodiment
It should be noted that in the case where not conflicting, the feature in embodiment and embodiment in the application can phase Mutually combine, the present invention is described in further detail with specific embodiment below in conjunction with the accompanying drawings.
Fig. 1 is the system flow chart of the future anticipation method of interaction medium in a kind of utilization dynamic scene of the invention. Mainly include data input;Multisample is generated;Sample sort with it is perfect;Scene is merged.
Wherein, multisample is generated, including universal model construction and condition variation self-encoding encoder.
Universal model is constructed, for playing interactive object in generation event in certain scene, no matter static or fortune Move, referred to as medium, if giving n medium, their passing movement track X=<X1,X2,…,Xn>, calculate theirs Future behaviour track Y=<Y1,Y2,…,Yn>Respective probability then be P<Y|X,I>, especially, I is defeated for the scene lower a moment Enter, the maximum probability obtained according to input is exported as prediction locus.
Described condition variation self-encoding encoder, learns a decision function f using depth generation model, and the function was incited somebody to action It is mapped to toward movement track X and input I on the Y of future behaviour track, specifically:
(1) random latent variable z is introducedi, in input XiUnder conditions of, study to output YiProbability distribution be P (Yi|Xi), It is respectively identification network Q that this process, which builds corresponding neutral net,φ(zi|Yi,Xi), condition pro-active network Pv(zi|Xi) and it is raw Into network Pθ(Yi|Xi,zi);
(2) in the training stage:1. respectively with two Recognition with Recurrent Neural Network to medium i, XiAnd YiEncoded, obtain two As a result 2. the two results are merged and is delivered to one layer of fully-connected network using nonlinear activation function again;3. again Generated with two parallel fully-connected networks on ziAverageAnd standard deviationGaussian distribution model is modeled as according to it, and And be normalized with KL divergences;4. propose that two loss functions are adjusted to self-encoding encoder, respectively rebuild loss function With KL divergence loss functions lKLD=DKL(Qφ(zi|Yi,Xi)‖Pv(zi));
(3) in test phase:1. due to future anticipation trackDo not apply to and be rejected, therefore utilize passing track With multiple latent variable zi (k)Sample merges;2. it is relative with the training stage, nowWith β (zi (k)) it is input to Recognition with Recurrent Neural Network solution The hypothesis set of multiple predictions is produced in code device.
Sample sort with it is perfect, including sample evaluation improve with iterative feedback two parts.
Sample evaluates perfect, is predicted for future event, using the decision-making method in enhancing learning framework, Jie Matter can have selected one after training rewards maximized action and goes to implement as following event for a long time, and now design one is followed Ring neural network model goes the prediction to being made in each trainingThe measurement of long term accumulation reward is carried out, is specially:
(1) medium i is scored, there is K forecast sampleThen the evaluation score s of each sample can Have:
Wherein,Refer to the forecast sample of other All Medias in addition to i,It is pre- test samples of the medium i in time t This,It is forecast samples all before knowing timestamp t, ψ is allocated to each time step t reward function, should Function is realized with the fully connected network network layers for being connected to Recognition with Recurrent Neural Network;
(2) perfect to medium i, in scoring process, Recognition with Recurrent Neural Network model estimates a regression vector simultaneouslyThe regression function η that the estimation procedure is used is:
Wherein, regression function η can accumulate all passing scene environments and the dynamic of all interaction mediums, while Estimated on whole time dimension optimal
Iterative feedback, optimum regression substituting vector is obtained using formula (2)Go gradually to improve the hypothesis set of predictionSpecially:
(1) each iteration,All it is updated toIt is then delivered to reverse optimum control module;
(2) reverse optimum control training in sequence and improve module, they have two kinds loss items, be respectively:Intersect Entropy loss lCE=H (p, q), wherein q are obtained by activation primitive, i.e.,Return Lose item
(3) to a training batch in neutral net, its whole multitask loss item is:
Wherein, N is the number of the batch medium.
Scene is merged, including context insertion and interaction feature.
Context is embedded in, using Recognition with Recurrent Neural Network, receives following input X on each timestamp tt
Wherein,It isIn the speed of t, γ is one layer of full articulamentum with nonlinear activation function, its handle Speed is mapped to higher-dimension and characterizes space,It will be located atConvolutional neural networks feature ρ (I) carry out pondization operate, R is gathered together the hiding vector on other medium spatial domains by a fused layer, self-embedding vectorIt is used as circulation nerve The initialization hidden state vector of network.
Interaction feature, using the spatial domain grid operations using pond layer operation, makes the phase between the different samples of different medium Interaction is unlikely to disappear, and is specially:In moment t medium i, its sample k spatial domain grid cell withCentered on, then to every One grid cell g, the feature of its interior Hide AllAverage pondization operation will all be carried out.
Fig. 2 is the example of the future anticipation method of interaction medium in a kind of utilization dynamic scene of the invention.As schemed Show, it is observed that by different iterationses, model learning to feedback gradually strengthen, the route of prediction is also increasingly leaned on Nearly real route, that is, thick line of adding some points.
For those skilled in the art, the present invention is not restricted to the details of above-described embodiment, in the essence without departing substantially from the present invention In the case of refreshing and scope, the present invention can be realized with other concrete forms.In addition, those skilled in the art can be to this hair Bright to carry out various changes and modification without departing from the spirit and scope of the present invention, these improvement and modification also should be regarded as the present invention's Protection domain.Therefore, appended claims are intended to be construed to include preferred embodiment and fall into all changes of the scope of the invention More and modification.

Claims (10)

1. a kind of future anticipation method of interaction medium in utilization dynamic scene, it is characterised in that mainly including multisample Generate (one);Sample sorts and perfect (two);Scene merges (three).
2. (one) is generated based on the multisample described in claims 1, it is characterised in that become including universal model construction and condition Divide self-encoding encoder.
3. based on the universal model construction described in claims 2, it is characterised in that for being risen in certain scene in generation event Interactive object, no matter static or motion, referred to as medium, if giving n medium, their passing movement track X=< X1,X2,…,Xn>, calculates their future behaviour track Y=< Y1,Y2,…,Yn> respective probability is then P < Y | X, I >, especially, I are the input at the scene lower a moment, and the maximum probability obtained according to input is exported as prediction locus.
4. based on the condition variation self-encoding encoder described in claims 2, it is characterised in that learnt using depth generation model Passing movement track X and input I are mapped on the Y of future behaviour track by one decision function f, the function, specifically:
(1) random latent variable z is introducedi, in input XiUnder conditions of, study to output YiProbability distribution be P (Yi|Xi), this mistake It is respectively identification network Q that journey, which builds corresponding neutral net,φ(zi|Yi,Xi), condition pro-active network Pv(zi|Xi) and generation net Network Pθ(Yi|Xi,zi);
(2) in the training stage:1. respectively with two Recognition with Recurrent Neural Network to medium i, XiAnd YiEncoded, obtain two results 2. the two results are merged and is delivered to one layer of fully-connected network using nonlinear activation function again;3. again with simultaneously Two capable fully-connected networks are generated on ziAverageAnd standard deviationGaussian distribution model is modeled as according to it, and is used KL divergences are normalized;4. propose that two loss functions are adjusted to self-encoding encoder, respectively rebuild loss function With KL divergence loss functions
(3) in test phase:1. due to future anticipation trackDo not apply to and be rejected, therefore utilize passing trackWith it is many Individual latent variable zi (k)Sample merges;2. it is relative with the training stage, nowWith β (zi (k)) it is input to Recognition with Recurrent Neural Network decoder The middle hypothesis set for producing multiple predictions.
5. based on the sample sequence described in claims 1 and perfect (two), it is characterised in that improve and change including sample evaluation Generation feedback two parts.
6. evaluate perfect based on the sample described in claims 5, it is characterised in that predicted for future event, using enhancing Decision-making method in learning framework, a medium can have selected after training one reward for a long time it is maximized action go reality Apply as following event, now design a Recognition with Recurrent Neural Network model and go prediction to being made in each trainingCarry out The measurement of long term accumulation reward, be specially:
(1) medium i is scored, there is K forecast sampleThen the evaluation score s of each sample can have:
<mrow> <mi>s</mi> <mrow> <mo>(</mo> <msubsup> <mover> <mi>Y</mi> <mo>^</mo> </mover> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>;</mo> <mi>I</mi> <mo>,</mo> <mi>X</mi> <mo>,</mo> <msubsup> <mover> <mi>Y</mi> <mo>^</mo> </mover> <mrow> <mi>j</mi> <mo>\</mo> <mi>i</mi> </mrow> <mrow> <mo>(</mo> <mo>&amp;ForAll;</mo> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> </mrow> <mo>=</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>t</mi> <mo>=</mo> <mn>1</mn> </mrow> </munder> <mi>&amp;psi;</mi> <mrow> <mo>(</mo> <msubsup> <mover> <mi>y</mi> <mo>^</mo> </mover> <mrow> <mi>i</mi> <mo>,</mo> <mi>t</mi> </mrow> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>;</mo> <mi>I</mi> <mo>,</mo> <mi>X</mi> <mo>,</mo> <msubsup> <mover> <mi>Y</mi> <mo>^</mo> </mover> <mrow> <mi>&amp;tau;</mi> <mo>&lt;</mo> <mi>t</mi> </mrow> <mrow> <mo>(</mo> <mo>&amp;ForAll;</mo> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>)</mo> </mrow> </mrow>
Wherein,Refer to the forecast sample of other All Medias in addition to i,It is forecast samples of the medium i in time t,It is forecast samples all before knowing timestamp t, ψ is allocated to each time step t reward function, the function It is to be realized with the fully connected network network layers for being connected to Recognition with Recurrent Neural Network;
(2) perfect to medium i, in scoring process, Recognition with Recurrent Neural Network model estimates a regression vector simultaneouslyShould The regression function η that estimation procedure is used is:
<mrow> <mi>&amp;Delta;</mi> <msubsup> <mover> <mi>Y</mi> <mo>^</mo> </mover> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>=</mo> <mi>&amp;eta;</mi> <mrow> <mo>(</mo> <msubsup> <mover> <mi>Y</mi> <mo>^</mo> </mover> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>;</mo> <mi>I</mi> <mo>,</mo> <mi>X</mi> <mo>,</mo> <msubsup> <mover> <mi>Y</mi> <mo>^</mo> </mover> <mrow> <mi>j</mi> <mo>\</mo> <mi>i</mi> </mrow> <mrow> <mo>(</mo> <mo>&amp;ForAll;</mo> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>2</mn> <mo>)</mo> </mrow> </mrow>
Wherein, regression function η can accumulate all passing scene environments and the dynamic of all interaction mediums, while whole Estimated on time dimension optimal
7. based on the iterative feedback described in claims 5, it is characterised in that using formula (2) obtain optimum regression replace to AmountGo gradually to improve the hypothesis set of predictionSpecially:
(1) each iteration,All it is updated toIt is then delivered to reverse optimum control module;
(2) reverse optimum control training in sequence and improve module, they have two kinds loss items, be respectively:Cross entropy is damaged Lose itemWherein q is obtained by activation primitive, i.e.,Return loss
(3) to a training batch in neutral net, its whole multitask loss item is:
Wherein, N is the number of the batch medium.
8. (three) are merged based on the scene described in claims 1, it is characterised in that including context insertion and interaction feature.
9. based on the context insertion described in claims 8, it is characterised in that Recognition with Recurrent Neural Network is used, at each Between receive following input X on stamp tt
<mrow> <msub> <mi>X</mi> <mi>t</mi> </msub> <mo>=</mo> <mo>&amp;lsqb;</mo> <mi>&amp;gamma;</mi> <mrow> <mo>(</mo> <msub> <mover> <mi>&amp;upsi;</mi> <mo>^</mo> </mover> <mrow> <mi>i</mi> <mo>,</mo> <mi>t</mi> </mrow> </msub> <mo>)</mo> </mrow> <mo>,</mo> <mi>p</mi> <mrow> <mo>(</mo> <msub> <mover> <mi>y</mi> <mo>^</mo> </mover> <mrow> <mi>i</mi> <mo>,</mo> <mi>t</mi> </mrow> </msub> <mo>;</mo> <mi>&amp;rho;</mi> <mo>(</mo> <mi>I</mi> <mo>)</mo> <mo>)</mo> </mrow> <mo>,</mo> <mi>r</mi> <mrow> <mo>(</mo> <msub> <mover> <mi>y</mi> <mo>^</mo> </mover> <mrow> <mi>i</mi> <mo>,</mo> <mi>t</mi> </mrow> </msub> <mo>;</mo> <msub> <mover> <mi>y</mi> <mo>^</mo> </mover> <mrow> <mi>j</mi> <mo>\</mo> <mi>i</mi> </mrow> </msub> <mo>,</mo> <msub> <mi>h</mi> <msub> <mover> <mi>Y</mi> <mo>^</mo> </mover> <mrow> <mi>j</mi> <mo>\</mo> <mi>i</mi> </mrow> </msub> </msub> <mo>)</mo> </mrow> <mo>&amp;rsqb;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>4</mn> <mo>)</mo> </mrow> </mrow>
Wherein,It isIn the speed of t, γ is one layer of full articulamentum with nonlinear activation function, and it reflects speed It is mapped to higher-dimension and characterizes space,It will be located atConvolutional neural networks feature ρ (I) carry out pondization and operate, r passes through One fused layer, the hiding vector on other medium spatial domains is gathered together, self-embedding vectorIt is used as Recognition with Recurrent Neural Network Initialize hidden state vector.
10. based on the interaction feature described in claims 8, it is characterised in that use the spatial domain grid using pond layer operation Operation, makes the interaction between the different samples of different medium is unlikely to disappear, is specially:In moment t medium i, its sample k's Spatial domain grid cell withCentered on, then to each grid cell g, the feature of its interior Hide All Average pondization operation will all be carried out.
CN201710487508.2A 2017-06-23 2017-06-23 A kind of future anticipation method of interaction medium in utilization dynamic scene Withdrawn CN107274029A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710487508.2A CN107274029A (en) 2017-06-23 2017-06-23 A kind of future anticipation method of interaction medium in utilization dynamic scene

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710487508.2A CN107274029A (en) 2017-06-23 2017-06-23 A kind of future anticipation method of interaction medium in utilization dynamic scene

Publications (1)

Publication Number Publication Date
CN107274029A true CN107274029A (en) 2017-10-20

Family

ID=60069234

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710487508.2A Withdrawn CN107274029A (en) 2017-06-23 2017-06-23 A kind of future anticipation method of interaction medium in utilization dynamic scene

Country Status (1)

Country Link
CN (1) CN107274029A (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108777140A (en) * 2018-04-27 2018-11-09 南京邮电大学 Phonetics transfer method based on VAE under a kind of training of non-parallel corpus
CN109472412A (en) * 2018-11-09 2019-03-15 百度在线网络技术(北京)有限公司 A kind of prediction technique and device of event
CN109710507A (en) * 2017-10-26 2019-05-03 北京京东尚科信息技术有限公司 A kind of method and apparatus of automatic test
CN110414989A (en) * 2019-07-29 2019-11-05 中国工商银行股份有限公司 Method for detecting abnormality and device, electronic equipment and computer readable storage medium
CN110929644A (en) * 2019-11-22 2020-03-27 南京甄视智能科技有限公司 Heuristic algorithm-based multi-model fusion face recognition method and device, computer system and readable medium
CN111563548A (en) * 2020-04-30 2020-08-21 鹏城实验室 Data preprocessing method and system based on reinforcement learning and related equipment
CN111599431A (en) * 2020-03-31 2020-08-28 太原金域临床检验有限公司 Report sheet-based data coding model generation method, system and equipment
CN111626472A (en) * 2020-04-17 2020-09-04 国网甘肃省电力公司信息通信公司 Scene trend judgment index computing system and method based on deep hybrid cloud model
CN111954860A (en) * 2018-01-21 2020-11-17 斯塔特斯公司 System and method for predicting fine-grained antagonistic multi-player movements
CN112418421A (en) * 2020-11-06 2021-02-26 常州大学 Roadside end pedestrian trajectory prediction algorithm based on graph attention self-coding model
CN113015983A (en) * 2018-10-24 2021-06-22 Hrl实验室有限责任公司 Autonomous system including continuous learning world model and related methods

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
NAMHOON LEE等: ""DESIRE: Distant Future Prediction in Dynamic Scenes with Interacting Agents"", 《网页在线公开:HTTP://EXPORT.ARXIV.ORG/ABS/1704.04394》 *

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109710507A (en) * 2017-10-26 2019-05-03 北京京东尚科信息技术有限公司 A kind of method and apparatus of automatic test
CN109710507B (en) * 2017-10-26 2022-03-04 北京京东尚科信息技术有限公司 Automatic testing method and device
CN111954860A (en) * 2018-01-21 2020-11-17 斯塔特斯公司 System and method for predicting fine-grained antagonistic multi-player movements
CN108777140A (en) * 2018-04-27 2018-11-09 南京邮电大学 Phonetics transfer method based on VAE under a kind of training of non-parallel corpus
CN113015983A (en) * 2018-10-24 2021-06-22 Hrl实验室有限责任公司 Autonomous system including continuous learning world model and related methods
CN109472412A (en) * 2018-11-09 2019-03-15 百度在线网络技术(北京)有限公司 A kind of prediction technique and device of event
CN110414989A (en) * 2019-07-29 2019-11-05 中国工商银行股份有限公司 Method for detecting abnormality and device, electronic equipment and computer readable storage medium
CN110929644A (en) * 2019-11-22 2020-03-27 南京甄视智能科技有限公司 Heuristic algorithm-based multi-model fusion face recognition method and device, computer system and readable medium
CN111599431A (en) * 2020-03-31 2020-08-28 太原金域临床检验有限公司 Report sheet-based data coding model generation method, system and equipment
CN111626472A (en) * 2020-04-17 2020-09-04 国网甘肃省电力公司信息通信公司 Scene trend judgment index computing system and method based on deep hybrid cloud model
CN111626472B (en) * 2020-04-17 2023-10-27 国网甘肃省电力公司信息通信公司 Scene trend judgment index computing system and method based on depth hybrid cloud model
CN111563548A (en) * 2020-04-30 2020-08-21 鹏城实验室 Data preprocessing method and system based on reinforcement learning and related equipment
CN111563548B (en) * 2020-04-30 2024-02-02 鹏城实验室 Data preprocessing method, system and related equipment based on reinforcement learning
CN112418421A (en) * 2020-11-06 2021-02-26 常州大学 Roadside end pedestrian trajectory prediction algorithm based on graph attention self-coding model
CN112418421B (en) * 2020-11-06 2024-01-23 常州大学 Road side end pedestrian track prediction algorithm based on graph attention self-coding model

Similar Documents

Publication Publication Date Title
CN107274029A (en) A kind of future anticipation method of interaction medium in utilization dynamic scene
Mousavi et al. Traffic light control using deep policy‐gradient and value‐function‐based reinforcement learning
CN109635917B (en) Multi-agent cooperation decision and training method
Bhattacharyya et al. Multi-agent imitation learning for driving simulation
Christiano et al. Supervising strong learners by amplifying weak experts
Castillo et al. Hybrid intelligent systems for time series prediction using neural networks, fuzzy logic, and fractal theory
Papageorgiou et al. A review of fuzzy cognitive maps research during the last decade
Papageorgiou Review study on fuzzy cognitive maps and their applications during the last decade
EP3602402B1 (en) Learning visual concepts using neural networks
Foster et al. Structure in the space of value functions
Papageorgiou Review study on fuzzy cognitive maps and their applications during the last decade
CN110147456A (en) A kind of image classification method, device, readable storage medium storing program for executing and terminal device
CN110991027A (en) Robot simulation learning method based on virtual scene training
Ghaderi et al. Behavioral simulation and optimization of generation companies in electricity markets by fuzzy cognitive map
Ulhaq et al. Efficient diffusion models for vision: A survey
Zheng et al. Synthetic dynamic PMU data generation: A generative adversarial network approach
Groumpos Modelling business and management systems using fuzzy cognitive maps: A critical overview
Rabe et al. Combining a discrete-event simulation model of a logistics network with deep reinforcement learning
Hu et al. Adaptive exploration strategy with multi-attribute decision-making for reinforcement learning
Andersen et al. The dreaming variational autoencoder for reinforcement learning environments
Daniel et al. A perceptually-validated metric for crowd trajectory quality evaluation
CN114116995A (en) Session recommendation method, system and medium based on enhanced graph neural network
CN117271768A (en) False news detection method and device based on large language model analysis and guidance
Sayama et al. Beyond social fragmentation: Coexistence of cultural diversity and structural connectivity is possible with social constituent diversity
Shafik et al. A reawakening of machine learning application in unmanned aerial vehicle: future research motivation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication
WW01 Invention patent application withdrawn after publication

Application publication date: 20171020