WO2016206491A1 - 目标对象运动轨迹确定方法、装置以及存储介质 - Google Patents
目标对象运动轨迹确定方法、装置以及存储介质 Download PDFInfo
- Publication number
- WO2016206491A1 WO2016206491A1 PCT/CN2016/081694 CN2016081694W WO2016206491A1 WO 2016206491 A1 WO2016206491 A1 WO 2016206491A1 CN 2016081694 W CN2016081694 W CN 2016081694W WO 2016206491 A1 WO2016206491 A1 WO 2016206491A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- control point
- training
- direction vector
- vector
- determining
- Prior art date
Links
- 230000033001 locomotion Effects 0.000 title claims abstract description 212
- 238000000034 method Methods 0.000 title claims abstract description 39
- 238000003860 storage Methods 0.000 title claims abstract description 17
- 238000012549 training Methods 0.000 claims description 309
- 238000004891 communication Methods 0.000 description 16
- 230000006870 function Effects 0.000 description 11
- 238000010586 diagram Methods 0.000 description 10
- 238000012545 processing Methods 0.000 description 6
- 238000001514 detection method Methods 0.000 description 4
- 230000008859 change Effects 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 2
- 230000005484 gravity Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000005236 sound signal Effects 0.000 description 2
- 241000699666 Mus <mouse, genus> Species 0.000 description 1
- 241000699670 Mus sp. Species 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000007599 discharging Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000007667 floating Methods 0.000 description 1
- 238000012905 input function Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 230000014759 maintenance of location Effects 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000010079 rubber tapping Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000010897 surface acoustic wave method Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/55—Controlling game characters or game objects based on the game progress
- A63F13/57—Simulating properties, behaviour or motion of objects in the game world, e.g. computing tyre load in a car race game
- A63F13/573—Simulating properties, behaviour or motion of objects in the game world, e.g. computing tyre load in a car race game using trajectories of game objects, e.g. of a golf ball according to the point of impact
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/20—Input arrangements for video game devices
- A63F13/21—Input arrangements for video game devices characterised by their sensors, purposes or types
- A63F13/214—Input arrangements for video game devices characterised by their sensors, purposes or types for locating contacts on a surface, e.g. floor mats or touch pads
- A63F13/2145—Input arrangements for video game devices characterised by their sensors, purposes or types for locating contacts on a surface, e.g. floor mats or touch pads the surface being also a display device, e.g. touch screens
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/20—Input arrangements for video game devices
- A63F13/21—Input arrangements for video game devices characterised by their sensors, purposes or types
- A63F13/215—Input arrangements for video game devices characterised by their sensors, purposes or types comprising means for detecting acoustic signals, e.g. using a microphone
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/25—Output arrangements for video game devices
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/30—Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers
- A63F13/35—Details of game servers
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/40—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
- A63F13/42—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
- A63F13/428—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle involving motion or position input signals, e.g. signals representing the rotation of an input controller or a player's arm motions sensed by accelerometers or gyroscopes
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/50—Controlling the output signals based on the game progress
- A63F13/52—Controlling the output signals based on the game progress involving aspects of the displayed game scene
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/20—Linear translation of whole images or parts thereof, e.g. panning
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/80—Special adaptations for executing a specific game genre or game mode
- A63F13/822—Strategy games; Role-playing games
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/60—Methods for processing data by generating or executing the game program
- A63F2300/64—Methods for processing data by generating or executing the game program for computing dynamical parameters of game objects, e.g. motion determination or computation of frictional forces for a virtual car
- A63F2300/646—Methods for processing data by generating or executing the game program for computing dynamical parameters of game objects, e.g. motion determination or computation of frictional forces for a virtual car for calculating the trajectory of an object
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30241—Trajectory
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H20/00—ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance
- G16H20/30—ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to physical therapies or activities, e.g. physiotherapy, acupressure or exercising
Definitions
- the present invention relates to the field of computer technologies, and in particular, to a method, an apparatus, and a storage medium for determining a target object motion trajectory.
- a role can be selected in the application, and the selected role can be referred to as a target object.
- the terminal may determine a motion trajectory of the target object on the motion track based on the user's control of the target object, thereby controlling the target object motion through the navigation module of the application.
- Embodiments of the present invention provide a method, an apparatus, and a storage medium for determining a target object motion trajectory.
- a method for determining a target object motion trajectory comprising:
- Determining motion of the target object on the motion trajectory by specifying a spline curve interpolation model based on the first control point, the second control point, the first direction vector, and the second direction vector Track.
- a target object motion trajectory determining apparatus comprising: one or more processors and a storage medium storing an operation instruction, when the operation instruction in the storage medium is executed, the processing is performed Perform the following steps:
- Determining motion of the target object on the motion trajectory by specifying a spline curve interpolation model based on the first control point, the second control point, the first direction vector, and the second direction vector Track.
- a non-transitory computer readable storage medium having stored thereon computer executable instructions for executing the executable instructions in a computer performs the following steps:
- Determining motion of the target object on the motion trajectory by specifying a spline curve interpolation model based on the first control point, the second control point, the first direction vector, and the second direction vector Track.
- the first direction vector and the second direction vector are the same, and the multiple variables in the specified spline interpolation model are offset, and the generated
- the actual motion trajectory of the target object is a straight line, not a curve, so that the actual motion trajectory of the target object is the same as the theoretical motion trajectory, which improves the accuracy of determining the motion trajectory of the target object.
- FIG. 1 is a flowchart of a method for determining a target object motion trajectory according to an embodiment of the present invention
- FIG. 2 is a flowchart of another method for determining a target object motion trajectory according to an embodiment of the present invention
- FIG. 3 is a schematic diagram of a line generated by a CatmullRom spline curve algorithm according to an embodiment of the present invention
- FIG. 4 is a schematic diagram of a target object motion track interface according to an embodiment of the present invention.
- FIG. 5 is a schematic diagram of another target object motion trajectory determining interface according to an embodiment of the present invention.
- FIG. 6 is a schematic diagram of a line generated by a specified spline curve interpolation model according to an embodiment of the present invention.
- FIG. 7 is a schematic structural diagram of a target object motion trajectory determining apparatus according to an embodiment of the present invention.
- FIG. 8 is a schematic structural diagram of another target object motion trajectory determining apparatus according to an embodiment of the present invention.
- FIG. 9 is a schematic structural diagram of still another target object motion trajectory determining apparatus according to an embodiment of the present invention.
- FIG. 1 is a flowchart of a method for determining a target object motion trajectory according to an embodiment of the present invention.
- the method can be applied to a terminal, which can be a mobile phone, a tablet computer, a palmtop computer, etc.
- the method includes steps 101 to 103.
- step 101 a first control point and a second control point on the motion track are acquired based on a current position of the target object on the motion track, and the first control point and the second control point are adjacent control points.
- step 102 a first direction vector and a second direction vector are obtained, the first direction vector is a unit direction vector at the first control point, and the second direction vector is a unit direction vector at the second control point.
- step 103 based on the first control point, the second control point, the first direction vector, and the second direction vector, the motion track of the target object on the motion track is determined by specifying a spline interpolation model.
- the first direction vector and the second direction vector are the same, and the multiple variables in the specified spline interpolation model are offset, and the generated The actual motion trajectory of the target object is a straight line, not a curve, and thus the target The actual motion trajectory of the object is the same as the theoretical motion trajectory, which improves the accuracy of determining the motion trajectory of the target object.
- the first control point, the second control point, the first direction vector, and the second direction vector may be included before the motion track of the target object on the motion track is determined by specifying a spline interpolation model. : obtaining a spline curve interpolation model to be trained;
- the order of the third training control point and the fourth training control point may be sequentially arranged;
- the specified spline interpolation model is determined based on the second training control point, the third training control point, the training distance, the first training vector, the second training vector, and the training spline interpolation model.
- determining the training distance, the first training vector, and the second training vector based on the first training control point, the second training control point, the third training control point, and the fourth training control point may include:
- a unit direction vector connecting the first training control point and the fourth training control point is determined as the second training vector.
- determining the specified spline interpolation model based on the second training control point, the third training control point, the training distance, the first training vector, the second training vector, and the training spline interpolation model may include:
- the specified spline interpolation model is determined.
- determining, according to the first control point, the second control point, the first direction vector, and the second direction vector, determining a motion track of the target object on the motion track by specifying a spline interpolation model may include:
- P(U) is the motion trajectory of the target object
- U is the interpolation ratio of the current motion distance between the first control point and the second control point
- P i-1 is the first control point
- P i is the second Control point
- ⁇ is the control point distance
- E is the first direction vector
- F is the second direction vector.
- the method may further include:
- the optional embodiments of the present invention may be used in any combination to form an optional embodiment of the present invention.
- FIG. 2 is a flowchart of a method for determining a target object motion trajectory according to an embodiment of the present invention.
- the method is applied to a terminal, which may be a mobile phone, a tablet computer, a palmtop computer, etc.
- the method includes steps 201 to 206.
- step 201 a training spline interpolation model is acquired.
- the generated line is a curve instead of a straight line by the CatmullRom spline curve algorithm. For example, as shown in FIG. 3, when the actual motion trajectory of the target object is a straight line from A to B. At the time, through the CatmullRom spline algorithm, the generated lines start from the line A to A', and then the line from A' to B.
- the training spline interpolation model is obtained.
- the training spline interpolation model can be:
- U is the interpolation ratio and is a floating point number of 0-1
- P(U) is the motion trajectory
- C 0 , C 1 , C 2 and C 3 are the parameters of the training spline curve interpolation model
- U specifies P(U) is a point on the motion trajectory. For example, when U is 0, P(0) is the starting point of the motion trajectory. When U is 1, P(1) is the end point of the motion trajectory. .
- step 202 based on the first training control point, the second training control point, the third training control point, and the fourth training control point, determining a training distance, a first training vector, and a second training vector, the first training control point,
- the second training control point, the third training control point, and the fourth training control point are control points sequentially arranged on the training track.
- a specified spline interpolation model can be derived based on the training motion trajectory. Obtaining a first training control point, a second training control point, a third training control point, and a fourth training control point, where the second training control point is a starting point of the training motion track, and the third training control point is an end point of the training motion track, The first training control point is a control point before the second training control point and adjacent to the second training control point, and the fourth training control point is after the third training control point, and the control adjacent to the third training control point point.
- the first training control point is P i-2
- the second training control point is P i-1
- the third training control point is P i
- the fourth training control point is P i+1
- the second training control point is The distance between the third training control points is ⁇ , that is, the training distance is ⁇
- the unit vector of the connection between the first training control point and the third training control point is E, that is, the first training The vector is E
- the unit vector of the connection between the first training control point and the fourth training control point is F, that is, the second training vector is F.
- a specified spline interpolation model is determined based on the second training control point, the third training control point, the training distance, the first training vector, the second training vector, and the training spline interpolation model.
- the interpolation model of the specified spline curve is determined based on each parameter of the training spline interpolation model and the interpolating model of the spline to be trained.
- P(0) is the starting point of the motion trajectory
- P(1) is the end point of the motion trajectory. Therefore, for the training trajectory between the second training control point and the third training control point, P(( 0) is the second training control point P i-1 , P(1) is the third training control point P i .
- the direction at the second training control point is a direction in which the connection between the third training control point and the first training control point is located
- the direction at the third training control point is a fourth training control point and a second training control The direction in which the connection between the points is located.
- P(U) is the training motion trajectory
- U is the interpolation ratio
- P i-1 is the starting point of the training motion trajectory, that is, the second training control point
- P i is the end point of the training motion trajectory
- ⁇ is the distance between the second training control point and the third training control point, that is, the training distance
- E is the first training vector
- F is the second training vector.
- the specified spline interpolation model can be trained by the above steps 201-203.
- the above steps 201-203 can be performed not only when determining the target object motion trajectory, but also by using the above steps 201-203 to obtain the specified sample.
- a curve interpolation model, and storing the specified spline interpolation model, and then determining the target object motion trajectory, the specified spline interpolation model can be directly obtained, and the timing of determining the specified spline interpolation model is not in the embodiment of the present invention. Make specific limits.
- a spline interpolation model can be specified to generate a spline of the motion track of the target object, and the spline can include multiple control points. Then, the terminal can also determine the motion trajectory of the target object on the motion track by the following steps, as follows.
- step 204 based on the current position of the target object on the motion track, the first control point and the second control point on the motion track are acquired, and the first control point and the second control point are adjacent control points.
- the terminal may determine the motion position of the target object on the motion track every specified duration, thereby obtaining a motion track of the target object on the motion track, where the specified duration is set in advance, for example, The duration can be 1 second, so based on the target object per second in the shipment
- the position of the motion on the moving orbit can determine that the motion of the target object on the moving orbit is relatively smooth.
- the speed of the target object on the motion track is constant, and therefore, when the target object starts moving from the start point of the motion track, every specified duration
- the current position of the target object on the motion track can be calculated based on the speed of the target object.
- the motion track may include a plurality of control points, and in the embodiment of the present invention, for each of the plurality of control points, the control point and the start point of the motion track may be calculated. The distance between the distances is divided by the total length of the motion track to obtain the length ratio of the control points, and the correspondence between the length ratio of each control point and each control point can be stored.
- the motion track includes 10 control points, the total length of the motion track is 100 meters, and from the first control point to the 10th control point, each control point is at the starting point of the motion track P 1 , 10 meters at P 2, 20 m P 3, 35 m P 4, 45 m P 5, 60 m P 6, 70 m P 7, 80 m P 8, 90 meters P 9, the end point P 10, That is, the distance between each control point and the starting point of the moving track is 0 meters, 10 meters, 20 meters, 35 meters, 45 meters, 60 meters, 70 meters, 80 meters, 90 meters, 100 meters, respectively.
- the length ratio of each control point is 0%, 10%, 20%, 35%, 45%, 60. %, 70%, 80%, 90%, 100%, and then the length ratio of each control point to each control point is stored in the correspondence between the control point and the length ratio shown in Table 1 below.
- the acquiring the first control point and the second control point on the motion track may include: obtaining a distance of the target object from the starting point according to the current position of the target object on the motion track, based on the current position of the target object on the motion track, The distance is divided by the total length of the motion track to obtain a length ratio of the current distance of the target object.
- the length ratio is adjacent to the two length ratios, and further, the control point corresponding to the smaller length ratio of the selected two length ratios can be determined as the first control point, and the larger length ratio of the selected two length ratios The corresponding control point is determined as the second control point.
- the distance of the target object from the starting point of the moving track is 50 meters
- the distance 50 meters is divided by the total length of the moving track by 100 meters
- the length ratio of the current moving distance is 50%.
- step 205 the first direction vector and the second direction vector are obtained, the first direction vector is a unit direction vector at the first control point, and the second direction vector is a unit direction vector at the second control point.
- the terminal may acquire the first direction vector and the second direction vector in multiple manners.
- a derivative operation can be performed on the specified spline interpolation model to obtain a derivative model of the specified spline interpolation model, and the interpolation ratio of the first control point is calculated, and the interpolation ratio of the first control point is substituted into the specified spline interpolation model.
- the derivative model obtains the first direction vector.
- the interpolation ratio of the second control point is calculated, and the interpolation ratio of the second control point is substituted into the derivative model of the specified spline interpolation model to obtain the second direction vector.
- the unit direction vector of the line between the second control point and the third control point can also be based on the coordinates of the second control point and the third Control point, calculate the unit direction vector of the line between the second control point and the third control point, and obtain the first direction vector, the third control point is before the first control point, and is adjacent to the first control point Control point; similarly, based on the coordinates of the fourth control point and the coordinates of the first control point, the unit direction vector of the connection between the fourth control point and the first control point can be calculated to obtain the second direction vector,
- the four control points are control points after the second control point and adjacent to the second control point. This embodiment of the present invention will not be listed one by one.
- a method for calculating an interpolation ratio of a first control point and an interpolation ratio of a second control point and an interpolation ratio method for calculating a current distance of the target object between a first control point and a second control point
- an interpolation ratio method for calculating a current distance of the target object between a first control point and a second control point For details, refer to step 206 for details. The embodiments of the present invention are not described in detail herein.
- the unit direction vector of the connection between the second control point and the third control point is calculated based on the coordinates of the second control point and the coordinates of the third control point
- the method for obtaining the first direction vector may include: The coordinates of the control point are subtracted from the coordinates of the third control point, the direction vector between the second control point and the third control point is obtained, and the direction vector between the second control point and the third control point is unitized to obtain the first A direction vector.
- the direction of calculating the second direction vector is similar, and the embodiment of the present invention will not elaborate on this.
- step 206 based on the first control point, the second control point, the first direction vector, and the second direction vector, the motion track of the target object on the motion track is determined by specifying a spline interpolation model.
- the terminal determines, according to the first control point and the second control point, an interpolation ratio between the first control point and the second control point of the target object; determining between the first control point and the second control point a distance obtained by the control point; and based on the first control point, the second control point, the current distance of the target object, the interpolation ratio between the first control point and the second control point, the control point distance, and the first direction vector And the second direction vector, determining the motion trajectory of the current motion of the target object by specifying the spline curve interpolation model as follows;
- P(U) in the above formula is the motion trajectory of the target object
- U is the interpolation ratio of the current motion distance of the target object between the first control point and the second control point
- P i-1 is the target object
- the starting point of the motion trajectory of the current motion that is, the first control point
- P i is the end point of the motion trajectory of the current motion of the target object, that is, the second control point
- ⁇ is between the first control point and the second control point
- E is the unit direction vector at the first control point, that is, the first direction vector
- F is the unit direction vector at the second control point, that is, the second direction vector.
- the specific operation of determining, by the terminal, the interpolation ratio of the current distance of the target object between the first control point and the second control point based on the first control point and the second control point may include: determining a current distance of the target object The length ratio is subtracted from the length ratio of the first control point to obtain a first ratio, and the length ratio of the second control point is subtracted from the length ratio of the first control point to obtain a second ratio, and the first ratio is divided by the second ratio. Obtaining an interpolation ratio of the current distance of the target object between the first control point and the second control point.
- control point distance can be calculated based on the coordinates of the first control point and the coordinates of the second control point.
- specific calculation method reference may be made to the related art, which is not described in detail in the embodiment of the present invention.
- the length ratio of the current distance of the target object is 50% minus the length ratio of the first control point P 5 by 45%, and the first ratio is 5%, and the length ratio of the second control point P 6 is 60%.
- the length ratio of the first control point P 5 is 45%, and the second ratio is 15%, and the first ratio 5% is divided by the second ratio 15%, and the current distance of the target object is obtained at the first control point P 5 and the first
- the interpolation ratio between the two control points P 6 is 33%.
- the distance between the first control point P 5 and the second control point P 6 is calculated to be 15 meters, that is, the control point distance is 15 meters, and then based on the first control point P 5 (0, 0,45), the second control point P 6 (0,0,60), the current moving distance of the target object is 33% of the interpolation ratio between the first control point and the second control point, and the control point distance is 15,
- the first direction vector (0, 0, 1) and the second direction vector (0, 0, 1) through the above specified spline interpolation model, determine the position of the current moving distance of the target object is (0, 0, 49.95), further determining the motion trajectory of the target object, that is, determining the position of the current moving distance of the target object from the first control point P 1 and the second control point P 2 in FIG. 4 (0, 0, 49.95)
- the specified spline interpolation model is a straight line. Therefore, no matter whether U is any value in 0-1, the phenomenon that the motion trajectory of the target object is a curve does not occur, and the accuracy of determining the motion trajectory is improved.
- the generated line is a line directly from A to B, and for the target moving object, it is also directly moved from A. To B, the fallback phenomenon shown in Figure 3 does not occur.
- the rate of change of the spline curve generated by the CatmullRom spline curve algorithm is changed, that is, the speed of the target object motion is also changed by the CatmullRom spline curve algorithm, as shown in the figure.
- the gap between the points from A' to B is changed from small to large, that is, for a constant specified duration, the speed at which the target object moves between A' and B is also slow. Fast changing.
- the speed of the target object movement needs to be constant, that is, the rate of change of the spline curve must be constant.
- the target object motion trajectory determining method may be applied to a scenario of a network game that interacts with an application server.
- the system for determining the motion trajectory may include a terminal and an application server, and the terminal and The application server is connected through a network, and the terminal can determine the motion track of the target object based on the foregoing method, and can synchronize the motion track of the target object to the application server, so that the application server performs the specified operation, for example, when the motion track reaches the motion track.
- the application server can reward the user corresponding to the terminal.
- the motion trajectory determining method can also be applied to other scenarios, for example, the scenario in which the terminal uses a network game in a single machine, that is, the system for determining the motion trajectory includes only the terminal.
- the embodiment of the present invention does not specifically limit the scenario for determining the motion track.
- the actual motion trajectory of the target object is a straight line
- the first direction vector and the second direction vector are the same
- the multiple variables in the specified spline interpolation model are offset
- the generated target is generated.
- the actual motion trajectory of the object is a straight line, not a curve, so that the actual motion trajectory of the target object is the same as the theoretical motion trajectory, which improves the accuracy of determining the motion trajectory of the target object.
- the velocity of the motion trajectory determined by the specified spline interpolation model is constant, and the phenomenon that the velocity of the target object is not constant does not occur.
- FIG. 7 is a schematic diagram of a target object motion trajectory determining apparatus according to an embodiment of the present invention.
- the apparatus includes:
- a first acquiring module 701 configured to acquire, according to a current position of the target object on the motion track, a first control point and a second control point on the motion track, where the first control point and the second control point are adjacent control points;
- the second obtaining module 702 is configured to obtain a first direction vector, where the first direction vector is a unit direction vector at the first control point, and a second direction vector is a unit direction vector at the second control point;
- the first determining module 703 is configured to determine a motion track of the target object on the motion track by specifying a spline interpolation model based on the first control point, the second control point, the first direction vector, and the second direction vector.
- the apparatus may further include:
- a third obtaining module 704 configured to acquire a training spline interpolation model
- a second determining module 705, configured to determine a training distance, a first training vector, and a second training vector based on the first training control point, the second training control point, the third training control point, and the fourth training control point, the first training
- the order of the control point, the second training control point, the third training control point, and the fourth training control point may be sequentially arranged;
- the third determining module 706 is configured to determine a specified spline interpolation model based on the second training control point, the third training control point, the training distance, the first training vector, the second training vector, and the training spline interpolation model.
- the second determining module 705 can include:
- a first determining unit configured to determine a distance between the second training control point and the third training control point Set as training distance
- a second determining unit configured to determine a unit direction vector of the connection between the first training control point and the third training control point as the first training vector
- a third determining unit configured to determine a unit direction vector of the connection between the first training control point and the fourth training control point as the second training vector.
- the third determining module 706 can include:
- a fourth determining unit configured to determine each parameter of the training spline curve interpolation model based on the second training control point, the third training control point, the training distance, the first training vector, and the second training vector;
- the fifth determining unit is configured to determine a specified spline interpolation model based on each parameter of the training spline interpolation model and the training spline interpolation model.
- the first determining module 703 may include:
- a sixth determining unit configured to determine, according to the first control point and the second control point, an interpolation ratio of a current moving distance of the target object on the motion track between the first control point and the second control point;
- a seventh determining unit configured to determine a distance between the first control point and the second control point, to obtain a control point distance
- An eighth determining unit configured to calculate an interpolation ratio, a control point distance, between the first control point and the second control point based on the first control point, the second control point, the current moving distance of the target object on the motion track,
- the first direction vector and the second direction vector determine the motion trajectory of the target object by specifying a spline interpolation model as follows;
- P(U) is the motion trajectory of the current motion of the target object
- U is the interpolation ratio of the current motion distance of the target object on the motion trajectory between the first control point and the second control point
- P i-1 is The first control point
- P i is the second control point
- ⁇ is the control point distance
- E is the first direction vector
- F is the second direction vector.
- the device may further include:
- the actual motion track of the target object is a straight line
- the multiple variables in the specified spline interpolation model are cancelled, and the actual motion trajectory of the generated target object is a straight line, not a curve, thereby causing the actual motion of the target object.
- the trajectory is the same as the theoretical trajectory, which improves the accuracy of determining the trajectory of the target object.
- the game motion trajectory determining apparatus provided in the above embodiment is only illustrated by the division of the above functional modules when the motion trajectory is determined. In actual applications, the functions may be assigned to different functional modules according to needs. Upon completion, the internal structure of the device is divided into different functional modules to perform all or part of the functions described above.
- the game motion trajectory determining apparatus and the game motion trajectory determining method embodiment are provided in the same concept, and the specific implementation process is described in detail in the method embodiment, and details are not described herein again.
- FIG. 9 is a structural block diagram of a target object motion trajectory determining apparatus according to an embodiment of the present invention.
- the apparatus may be a terminal, and the terminal 900 may include a communication unit 910 including one or more computer readable storages.
- WIFI Wireless Fidelity
- the terminal structure shown in FIG. 9 does not constitute a limitation to the terminal, and may include more or less components than those illustrated, or combine some components, or different component arrangements.
- the communication unit 910 can be used for transmitting and receiving information and receiving and transmitting signals during a call.
- the communication unit 910 can be an RF (Radio Frequency) circuit, a router, a modem, or the like.
- RF circuits as communication units include, but are not limited to, an antenna, at least one amplifier, a tuner, one or more oscillators, a Subscriber Identity Module (SIM) card, a transceiver, a coupler, and a LNA (Low Noise Amplifier, low).
- SIM Subscriber Identity Module
- the communication unit 910 can also communicate with the network and other devices through wireless communication.
- the wireless communication may use any communication standard or protocol, including but not limited to GSM (Global System of Mobile communication), GPRS (General Packet Radio Service), CDMA (Code Division Multiple Access). , code division multiple access), WCDMA (Wideband Code Division Multiple Access, LTE (Long Term Evolution), e-mail, SMS (Short Messaging Service), etc.
- the memory 920 can be used to store software programs and modules, and the processor 980 executes various functional applications and data processing by running software programs and modules stored in the memory 920.
- the memory 920 may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application required for at least one function (such as a sound playing function, an image playing function, etc.), and the like; the storage data area may be stored according to The data created by the use of the terminal 900 (such as audio data, phone book, etc.) and the like.
- memory 920 can include high speed random access memory, and can also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device. Accordingly, memory 920 may also include a memory controller to provide access to memory 920 by processor 980 and input unit 930.
- the input unit 930 can be configured to receive input numeric or character information and to generate keyboard, mouse, joystick, optical or trackball signal inputs related to user settings and function controls.
- the input unit 930 can include a touch sensitive surface 931 as well as other input devices 932.
- a touch-sensitive surface 931 also referred to as a touch display or trackpad, can collect touch operations on or near the user (eg, the user uses a finger, stylus, etc., any suitable object or accessory on the touch-sensitive surface 931 or The operation near the touch-sensitive surface 931) and drive the corresponding connecting device according to a preset program.
- the touch sensitive surface 931 can include two portions of a touch detection device and a touch controller.
- the touch detection device detects the touch orientation of the user, and detects a signal brought by the touch operation, and transmits the signal to the touch controller; the touch controller receives the touch information from the touch detection device, converts the touch information into contact coordinates, and sends the touch information.
- the processor 980 is provided and can receive commands from the processor 980 and execute them.
- the touch sensitive surface 931 can be implemented in various types such as resistive, capacitive, infrared, and surface acoustic waves.
- the input unit 930 can also include other input devices 932.
- other input devices 932 may include, but are not limited to, one or more of a physical keyboard, function keys (such as volume control buttons, switch buttons, etc.), trackballs, mice, joysticks, and the like.
- Display unit 940 can be used to display information entered by the user or information provided to the user and various graphical user interfaces of terminal 900, which can be constructed from graphics, text, icons, video, and any combination thereof.
- the display unit 940 can include a display panel 941.
- an LCD Liquid Crystal Display
- OLED Organic Light-Emitting
- the display panel 941 is configured in the form of a Diode, an organic light emitting diode, or the like.
- touch-sensitive surface 931 can cover the display panel 941, and when the touch-sensitive surface 931 detects a touch operation thereon or nearby, it is transmitted to the processor 980 to determine the type of the touch event, and then the processor 980 according to the touch event The type provides a corresponding visual output on display panel 941.
- touch-sensitive surface 931 and display panel 941 are implemented as two separate components to implement input and input functions, in some embodiments, touch-sensitive surface 931 can be integrated with display panel 941 for input. And output function.
- Terminal 900 can also include at least one type of sensor 950, such as a light sensor, motion sensor, and other sensors.
- the light sensor may include an ambient light sensor and a proximity sensor, wherein the ambient light sensor may adjust the brightness of the display panel 941 according to the brightness of the ambient light, and the proximity sensor may turn off the display panel 941 and/or the backlight when the terminal 900 moves to the ear.
- the gravity acceleration sensor can detect the magnitude of acceleration in all directions (usually three axes). When it is stationary, it can detect the magnitude and direction of gravity.
- the terminal 900 can also be configured with gyroscopes, barometers, hygrometers, thermometers, infrared sensors and other sensors, here Let me repeat.
- An audio circuit 960, a speaker 961, and a microphone 962 can provide an audio interface between the user and the terminal 900.
- the audio circuit 960 can transmit the converted electrical data of the received audio data to the speaker 961, and convert it into a sound signal output by the speaker 961.
- the microphone 962 converts the collected sound signal into an electrical signal, and the audio circuit 960 After receiving, it is converted to audio data, and then processed by the audio data output processor 980, transmitted to the other terminal, for example, via the communication unit 910, or outputted to the memory 920 for further processing.
- the audio circuit 960 may also include an earbud jack to provide communication of the peripheral earphones with the terminal 900.
- the terminal may be configured with a wireless communication unit 970, which may be a WIFI module.
- the WIFI is a short-range wireless transmission technology, and the terminal 900 can help the user to send and receive emails, browse web pages, and access streaming media through the wireless communication unit 970, which provides wireless broadband Internet access for users.
- the wireless communication unit 970 is shown in the drawing, it can be understood that it does not belong to the essential configuration of the terminal 900, and may be omitted as needed within the scope of not changing the essence of the invention.
- the processor 980 is a control center of the terminal 900, and connects the entire mobile phone by using various interfaces and lines.
- the various components perform overall monitoring of the handset by running or executing software programs and/or modules stored in memory 920, as well as invoking data stored in memory 920, performing various functions and processing data of terminal 900.
- the processor 980 may include one or more processing cores; preferably, the processor 980 may integrate an application processor and a modem processor, where the application processor mainly processes an operating system, a user interface, an application, and the like.
- the modem processor primarily handles wireless communications. It will be appreciated that the above described modem processor may also not be integrated into the processor 980.
- the terminal 900 also includes a power source 990 (such as a battery) that supplies power to the various components.
- a power source 990 such as a battery
- the power source can be logically coupled to the processor 980 through a power management system to manage functions such as charging, discharging, and power management through the power management system.
- Power supply 960 may also include any one or more of a DC or AC power source, a recharging system, a power failure detection circuit, a power converter or inverter, a power status indicator, and the like.
- the terminal 900 may further include a camera, a Bluetooth module, and the like, and details are not described herein again.
- the terminal further includes one or more programs, the one or more programs being stored in the memory and configured to be executed by one or more processors, the one or more programs including
- the instructions for performing the game motion trajectory determining method provided by the embodiment of the present invention include:
- the motion track of the target object on the motion track is determined by specifying a spline interpolation model.
- the method before determining the motion track of the target object on the motion track by specifying the spline interpolation model based on the first control point, the second control point, the first direction vector, and the second direction vector, the method further includes:
- the order of the third training control point and the fourth training control point may be sequentially arranged;
- the specified spline interpolation model is determined based on the second training control point, the third training control point, the training distance, the first training vector, the second training vector, and the training spline interpolation model.
- determining the training distance, the first training vector, and the second training vector based on the first training control point, the second training control point, the third training control point, and the fourth training control point may include:
- a unit direction vector connecting the first training control point and the fourth training control point is determined as the second training vector.
- determining the specified spline interpolation model based on the second training control point, the third training control point, the training distance, the first training vector, the second training vector, and the spline curve interpolation model to be trained may include:
- the specified spline interpolation model is determined.
- determining, according to the first control point, the second control point, the first direction vector, and the second direction vector, determining a motion track of the target object on the motion track by specifying a spline interpolation model may include:
- P(U) is the motion trajectory of the target object
- U is the interpolation ratio of the current motion distance between the first control point and the second control point
- P i-1 is the first control point
- P i is the second Control point
- ⁇ is the control point distance
- E is the first direction vector
- F is the second direction vector.
- the method may further include:
- the first direction vector and the second direction vector are the same, and the multiple variables in the specified spline interpolation model are offset, and the generated
- the actual motion trajectory of the target object is a straight line, not a curve, so that the actual motion trajectory of the target object is the same as the theoretical motion trajectory, which improves the accuracy of determining the motion trajectory of the target object.
- a person skilled in the art may understand that all or part of the steps of implementing the above embodiments may be completed by hardware, or may be instructed by a program to execute related hardware, and the program may be stored in a computer readable storage medium.
- the storage medium mentioned may be a read only memory, a magnetic disk or an optical disk or the like.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Manipulator (AREA)
- Processing Or Creating Images (AREA)
- Image Analysis (AREA)
Abstract
Description
控制点 | 长度比例 |
P1 | 0% |
P2 | 10% |
P3 | 20% |
P4 | 35% |
P5 | 45% |
P6 | 60% |
P7 | 70% |
P8 | 80% |
P9 | 90% |
P10 | 100% |
Claims (13)
- 一种目标对象运动轨迹确定方法,包括:基于目标对象在运动轨道上的当前位置,获取所述运动轨道上的第一控制点和第二控制点,所述第一控制点和所述第二控制点为相邻控制点;获取第一方向向量和第二方向向量,所述第一方向向量为所述第一控制点处的单位方向向量,所述第二方向向量为所述第二控制点处的单位方向向量;以及基于所述第一控制点、所述第二控制点、所述第一方向向量和所述第二方向向量,通过指定样条曲线插值模型,确定所述目标对象在所述运动轨道上的运动轨迹。
- 如权利要求1所述的方法,所述基于所述第一控制点、所述第二控制点、所述第一方向向量和所述第二方向向量,通过指定样条曲线插值模型,确定所述目标对象在所述运动轨道上的运动轨迹之前,还包括:获取训练样条曲线插值模型;基于第一训练控制点、第二训练控制点、第三训练控制点和第四训练控制点,确定训练距离、第一训练向量和第二训练向量,所述第一训练控制点、所述第二训练控制点、所述第三训练控制点和所述第四训练控制点为训练轨道上依次排列的控制点;以及基于所述第二训练控制点、所述第三训练控制点、所述训练距离、所述第一训练向量、所述第二训练向量和所述待训练样条曲线插值模型,确定所述指定样条曲线插值模型。
- 如权利要求2所述的方法,所述基于第一训练控制点、第二训练控制点、第三训练控制点和第四训练控制点,确定训练距离、第一训练向量和第二训练向量,包括:将所述第二训练控制点与所述第三训练控制点之间的距离,确定为所述训练距离;将所述第一训练控制点与所述第三训练控制点之间连线的单位方向向量, 确定为所述第一训练向量;以及将所述第一训练控制点与所述第四训练控制点之间连线的单位方向向量,确定为所述第二训练向量。
- 如权利要求2所述的方法,所述基于所述第二训练控制点、所述第三训练控制点、所述训练距离、所述第一训练向量、所述第二训练向量和所述待训练样条曲线插值模型,确定所述指定样条曲线插值模型,包括:基于所述第二训练控制点、所述第三训练控制点、所述训练距离、所述第一训练向量和所述第二训练向量,确定所述待训练样条曲线插值模型的参数;以及基于所述待训练样条曲线插值模型的参数和所述训练样条曲线插值模型,确定所述指定样条曲线插值模型。
- 如权利要求1-4任一权利要求所述的方法,所述基于所述第一控制点、所述第二控制点、所述第一方向向量和所述第二方向向量,通过指定样条曲线插值模型,确定所述目标对象在所述运动轨道上的运动轨迹,包括:基于所述第一控制点和所述第二控制点,确定所述目标物体的当前位置在所述第一控制点与所述第二控制点之间的插值比例;确定所述第一控制点与所述第二控制点之间的距离,得到控制点距离;以及基于所述第一控制点、所述第二控制点、所述插值比例、所述控制点距离、所述第一方向向量和所述第二方向向量,通过如下的指定样条曲线插值模型,确定所述目标对象当前运动的运动轨迹;P(U)=Pi-1+λEU+(3Pi-3Pi-1-λF-2λF)U2+(-2Pi+2Pi-1+λF+λE)U3其中,P(U)为所述目标对象的运动轨迹,U为所述插值比例,Pi-1为所述第一控制点,Pi为所述第二控制点,λ为所述控制点距离,E为所述第一方向向量,F为所述第二方向向量。
- 如权利要求1或5所述的方法,还包括:当所述第一方向向量与所述第二方向向量相等时,所述指定样条曲线插值模型为:P(U)=Pi-1+λVU,所述目标对象在所述运动轨道上的运动轨迹为直线,其中,V为所述第一方向向量或者所述第二方向向量。
- 一种目标对象运动轨迹确定装置,包括:一个或多个处理器和存储有操作指令的存储介质,当运行所述存储介质中的操作指令时,所述处理器执行如下步骤:基于目标对象在运动轨道上的当前位置,获取所述运动轨道上的第一控制点和第二控制点,所述第一控制点和所述第二控制点为相邻控制点;获取第一方向向量和第二方向向量,所述第一方向向量为所述第一控制点处的单位方向向量,所述第二方向向量为所述第二控制点处的单位方向向量;以及基于所述第一控制点、所述第二控制点、所述第一方向向量和所述第二方向向量,通过指定样条曲线插值模型,确定所述目标对象在所述运动轨道上的运动轨迹。
- 如权利要求7所述的装置,所述处理器还执行:获取训练样条曲线差值模型;基于第一训练控制点、第二训练控制点、第三训练控制点和第四训练控制点,确定训练距离、第一训练向量和第二训练向量,所述第一训练控制点、所述第二训练控制点、所述第三训练控制点和所述第四训练控制点为训练轨道上依次排列的控制点;以及基于所述第二训练控制点、所述第三训练控制点、所述训练距离、所述第一训练向量、所述第二训练向量和所述待训练样条曲线插值模型,确定所述指定样条曲线插值模型。
- 如权利要求8所述的装置,当所述处理器执行基于第一训练控制点、第二训练控制点、第三训练控制点和第四训练控制点,确定训练距离、第一训练向量和第二训练向量时,所述处理器执行:将所述第二训练控制点与所述第三训练控制点之间的距离,确定为所述训练距离;将所述第一训练控制点与所述第三训练控制点之间连线的单位方向向量,确定为所述第一训练向量;以及将所述第一训练控制点与所述第四训练控制点之间连线的单位方向向量,确定为所述第二训练向量。
- 如权利要求8所述的装置,当所述处理器执行基于所述第二训练控制点、所述第三训练控制点、所述训练距离、所述第一训练向量、所述第二训练向量和所述待训练样条曲线插值模型,确定所述指定样条曲线插值模型时,所述处理器执行:基于所述第二训练控制点、所述第三训练控制点、所述训练距离、所述第一训练向量和所述第二训练向量,确定所述待训练样条曲线插值模型的参数;以及基于所述待训练样条曲线插值模型的参数和所述训练样条曲线插值模型,确定所述指定样条曲线插值模型。
- 如权利要求7-10任一权利要求所述的装置,所述处理器执行基于所述第一控制点、所述第二控制点、所述第一方向向量和所述第二方向向量,通过指定样条曲线插值模型,确定所述目标对象在所述运动轨道上的运动轨迹时,所述处理器执行:基于所述第一控制点和所述第二控制点,确定所述目标物体的当前位置在所述第一控制点与所述第二控制点之间的插值比例;确定所述第一控制点与所述第二控制点之间的距离,得到控制点距离;以及基于所述第一控制点、所述第二控制点、所述插值比例、所述控制点距离、所述第一方向向量和所述第二方向向量,通过如下的指定样条曲线插值模型,确定所述目标对象当前运动的运动轨迹;P(U)=Pi-1+λEU+(3Pi-3Pi-1-λF-2λF)U2+(-2Pi+2Pi-1+λF+λE)U3其中,P(U)为所述目标对象的运动轨迹,U为所述插值比例,Pi-1为所述第一控制点,Pi为所述第二控制点,λ为所述控制点距离,E为所述第一方向向量,F为所述第二方向向量。
- 如权利要求7或11所述的装置,当所述第一方向向量与所述第二方向向量相等时,所述指定样条曲线插值模型为:P(U)=Pi-1+λVU,所述目标对象在所述运动轨道上的运动轨迹为直线,其中,V为所述第一方向向量或者所述第二方向向量。
- 一种非瞬时性的计算机可读存储介质,其上存储有计算机可执行指令,当计算机中运行这些可执行指令时,执行如下步骤:基于目标对象在运动轨道上的当前位置,获取获取所述运动轨道上的第一控制点和第二控制点,所述第一控制点和所述第二控制点为相邻控制点;获取第一方向向量和第二方向向量,所述第一方向向量为所述第一控制点处的单位方向向量,所述第二方向向量为所述第二控制点处的单位方向向量;以及基于所述第一控制点、所述第二控制点、所述第一方向向量和所述第二方向向量,通过指定样条曲线插值模型,确定所述目标对象在所述运动轨道上的运动轨迹。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020177016338A KR101975689B1 (ko) | 2015-06-24 | 2016-05-11 | 타겟 대상의 모션 궤적을 결정하는 방법 및 디바이스, 및 저장 매체 |
JP2017538656A JP6735760B2 (ja) | 2015-06-24 | 2016-05-11 | ターゲット対象の動き軌道を決定するための方法およびデバイス、ならびに記憶媒体 |
US15/624,245 US10354393B2 (en) | 2015-06-24 | 2017-06-15 | Method and device for determining motion trajectory of target subject, and storage medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510354868.6A CN105046059B (zh) | 2015-06-24 | 2015-06-24 | 游戏运动轨迹确定方法及装置 |
CN201510354868.6 | 2015-06-24 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/624,245 Continuation US10354393B2 (en) | 2015-06-24 | 2017-06-15 | Method and device for determining motion trajectory of target subject, and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2016206491A1 true WO2016206491A1 (zh) | 2016-12-29 |
Family
ID=54452599
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2016/081694 WO2016206491A1 (zh) | 2015-06-24 | 2016-05-11 | 目标对象运动轨迹确定方法、装置以及存储介质 |
Country Status (5)
Country | Link |
---|---|
US (1) | US10354393B2 (zh) |
JP (1) | JP6735760B2 (zh) |
KR (1) | KR101975689B1 (zh) |
CN (1) | CN105046059B (zh) |
WO (1) | WO2016206491A1 (zh) |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105046059B (zh) * | 2015-06-24 | 2017-09-29 | 深圳市腾讯计算机系统有限公司 | 游戏运动轨迹确定方法及装置 |
JP6429217B2 (ja) * | 2015-07-22 | 2018-11-28 | 本田技研工業株式会社 | 経路生成装置、経路生成方法、および経路生成プログラム |
CN107019915B (zh) * | 2016-02-01 | 2018-09-07 | 腾讯科技(深圳)有限公司 | 一种确定移动轨迹的方法、用户设备及系统 |
CN106055245A (zh) * | 2016-05-20 | 2016-10-26 | 小天才科技有限公司 | 一种计算游戏物体运动轨迹的方法及装置 |
CN108269297B (zh) * | 2017-12-27 | 2021-06-01 | 福建省天奕网络科技有限公司 | 一种在三维场景中编排角色运动轨迹的方法及终端 |
CN109126120B (zh) * | 2018-08-17 | 2022-06-03 | Oppo广东移动通信有限公司 | 马达控制方法及相关产品 |
CN111311745B (zh) * | 2018-12-11 | 2023-06-13 | 网易(杭州)网络有限公司 | 一种模型的放置方法和装置 |
SE544061C2 (en) * | 2019-07-05 | 2021-11-30 | Climeon Ab | Method and controller for dynamically determining a system curve in a heat power system |
CN112802067B (zh) * | 2021-01-26 | 2024-01-26 | 深圳市普汇智联科技有限公司 | 一种基于图网络的多目标跟踪方法及系统 |
CN113626118B (zh) * | 2021-07-30 | 2023-07-25 | 中汽创智科技有限公司 | 能耗实时显示方法、装置及设备 |
CN115131471B (zh) * | 2022-08-05 | 2024-08-02 | 北京字跳网络技术有限公司 | 基于图像的动画生成方法、装置、设备及存储介质 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080262721A1 (en) * | 2007-04-17 | 2008-10-23 | Hitachi, Ltd. | Map generation system and map generation method by using GPS tracks |
CN102794767A (zh) * | 2012-08-31 | 2012-11-28 | 江南大学 | 视觉引导的机器人关节空间b样条轨迹规划方法 |
CN103793933A (zh) * | 2012-11-02 | 2014-05-14 | 同济大学 | 虚拟人体动画的运动路径生成方法 |
CN103902086A (zh) * | 2012-12-28 | 2014-07-02 | 北京汇冠新技术股份有限公司 | 一种基于曲线拟合的触摸轨迹平滑方法及系统 |
CN105046059A (zh) * | 2015-06-24 | 2015-11-11 | 深圳市腾讯计算机系统有限公司 | 游戏运动轨迹确定方法及装置 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6115051A (en) | 1996-08-07 | 2000-09-05 | Adobe Systems Incorporated | Arc-length reparameterization |
US7805442B1 (en) * | 2000-12-05 | 2010-09-28 | Navteq North America, Llc | Method and system for representation of geographical features in a computer-based system |
US7477988B2 (en) | 2006-05-16 | 2009-01-13 | Navteq North America, Llc | Dual road geometry representation for position and curvature-heading |
JP5446136B2 (ja) * | 2008-06-05 | 2014-03-19 | 日本電気株式会社 | 走行支援システムおよび走行路曲線生成方法 |
CN101822545B (zh) * | 2010-05-11 | 2011-05-25 | 河南大学 | 一种数字减影血管造影运动伪影消除方法及其系统 |
CN102467587B (zh) * | 2010-11-01 | 2014-01-29 | 财团法人工业技术研究院 | 水冷机动态特性模型建立方法、水冷机监控方法和装置 |
CN102201122B (zh) * | 2011-05-16 | 2014-04-23 | 大连大学 | 一种运动捕捉的数据降噪方法、系统及运动捕捉系统 |
US9727987B2 (en) * | 2014-05-12 | 2017-08-08 | Adobe Systems Incorporated | Blending techniques for curve fitting |
-
2015
- 2015-06-24 CN CN201510354868.6A patent/CN105046059B/zh active Active
-
2016
- 2016-05-11 WO PCT/CN2016/081694 patent/WO2016206491A1/zh active Application Filing
- 2016-05-11 KR KR1020177016338A patent/KR101975689B1/ko active IP Right Grant
- 2016-05-11 JP JP2017538656A patent/JP6735760B2/ja active Active
-
2017
- 2017-06-15 US US15/624,245 patent/US10354393B2/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080262721A1 (en) * | 2007-04-17 | 2008-10-23 | Hitachi, Ltd. | Map generation system and map generation method by using GPS tracks |
CN102794767A (zh) * | 2012-08-31 | 2012-11-28 | 江南大学 | 视觉引导的机器人关节空间b样条轨迹规划方法 |
CN103793933A (zh) * | 2012-11-02 | 2014-05-14 | 同济大学 | 虚拟人体动画的运动路径生成方法 |
CN103902086A (zh) * | 2012-12-28 | 2014-07-02 | 北京汇冠新技术股份有限公司 | 一种基于曲线拟合的触摸轨迹平滑方法及系统 |
CN105046059A (zh) * | 2015-06-24 | 2015-11-11 | 深圳市腾讯计算机系统有限公司 | 游戏运动轨迹确定方法及装置 |
Also Published As
Publication number | Publication date |
---|---|
KR101975689B1 (ko) | 2019-05-07 |
JP6735760B2 (ja) | 2020-08-05 |
US10354393B2 (en) | 2019-07-16 |
JP2018506118A (ja) | 2018-03-01 |
US20170287142A1 (en) | 2017-10-05 |
CN105046059B (zh) | 2017-09-29 |
KR20170086572A (ko) | 2017-07-26 |
CN105046059A (zh) | 2015-11-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2016206491A1 (zh) | 目标对象运动轨迹确定方法、装置以及存储介质 | |
JP6467526B2 (ja) | 通信メッセージ送信方法及びウェアラブル・デバイス | |
WO2016169465A1 (zh) | 一种显示弹幕信息的方法、装置和系统 | |
WO2016197758A1 (zh) | 信息推荐系统、方法及装置 | |
WO2017041664A1 (zh) | 一种征信评分确定方法、装置及存储介质 | |
WO2016107501A1 (zh) | 智能设备控制方法及装置 | |
WO2017125027A1 (zh) | 一种进行信息展示的方法和装置、计算机存储介质 | |
CN104967896A (zh) | 一种显示弹幕评论信息的方法和装置 | |
US10845981B2 (en) | Operation control method, device and storage medium | |
CN106487984B (zh) | 一种调整音量的方法和装置 | |
WO2014000656A1 (zh) | 刷新页面的方法、装置及终端 | |
CN108984066B (zh) | 一种应用程序图标显示方法及移动终端 | |
TW201515682A (zh) | 一種數據獲取的方法及終端 | |
CN110874128B (zh) | 可视化数据处理方法和电子设备 | |
CN105022552A (zh) | 一种显示消息列表的方法和装置 | |
US20190034078A1 (en) | Method for Sliding Response Acceleration and Related Products | |
JP2017509051A (ja) | ストリーミングメディアデータに関する統計を収集するための方法およびシステム、ならびに関連する装置 | |
WO2017128986A1 (zh) | 多媒体菜单项的选择方法、装置及存储介质 | |
WO2015180596A1 (en) | Method for superposing location information on collage, terminal and server | |
CN111651030B (zh) | 传感器检测方法、装置、存储介质及移动终端 | |
US20160119695A1 (en) | Method, apparatus, and system for sending and playing multimedia information | |
WO2018219118A1 (zh) | 界面显示方法及相关产品 | |
CN107193551B (zh) | 一种生成图像帧的方法和装置 | |
CN105159655B (zh) | 行为事件的播放方法和装置 | |
WO2013152656A1 (zh) | 一种绘制滑动轨迹的方法及移动终端 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 16813607 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 20177016338 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2017538656 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC, EPO FORM 1205A DATED 30.05.18 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 16813607 Country of ref document: EP Kind code of ref document: A1 |