WO2023142353A1 - Pose prediction method and apparatus - Google Patents

Pose prediction method and apparatus Download PDF

Info

Publication number
WO2023142353A1
WO2023142353A1 PCT/CN2022/100638 CN2022100638W WO2023142353A1 WO 2023142353 A1 WO2023142353 A1 WO 2023142353A1 CN 2022100638 W CN2022100638 W CN 2022100638W WO 2023142353 A1 WO2023142353 A1 WO 2023142353A1
Authority
WO
WIPO (PCT)
Prior art keywords
target device
predicted
pose
target
visual information
Prior art date
Application number
PCT/CN2022/100638
Other languages
French (fr)
Chinese (zh)
Inventor
陈星鑫
庞敏健
万培佩
刘贤焯
Original Assignee
奥比中光科技集团股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 奥比中光科技集团股份有限公司 filed Critical 奥比中光科技集团股份有限公司
Publication of WO2023142353A1 publication Critical patent/WO2023142353A1/en

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/20Instruments for performing navigational calculations
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/3407Route searching; Route guidance specially adapted for specific applications
    • G01C21/343Calculating itineraries, i.e. routes leading from a starting point to a series of categorical destinations using a global route restraint, round trips, touristic trips
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"

Definitions

  • the present application relates to the technical field of positioning, and in particular to a pose prediction method and device.
  • SLAM Simultaneous Localization and Mapping
  • SLAM is a method that uses its own sensors to perceive the environment, calculate its own pose in real time, and build incremental maps.
  • Technology it does not need to modify the external environment, and its positioning accuracy can reach centimeter level. Its application fields can include AR/VR, robots, unmanned driving, drones, etc.
  • the positioning of the existing SLAM system is very dependent on visual information. If the environment texture is weak and the equipment moves faster within a certain period of time, the image in the visual information collected by the SLAM system will appear as weak texture, occlusion or Blurred, which affects the normal operation of the SLAM system. For example, when the visual information is not collected, the SLAM system cannot estimate the pose of the device, which makes the SLAM system unable to work and cannot be restarted, and must wait until the visual information is re-collected. to restart and work. Therefore, there is an urgent need for a technical solution that can solve the problem that the SLAM system cannot work and cannot be restarted because the SLAM system cannot estimate the pose of the device when the visual information is not collected.
  • the embodiment of the present application provides a pose prediction method, device, computer equipment, and computer-readable storage medium to solve the problem that the SLAM system cannot perform the pose prediction of the device when no visual information is collected in the prior art. It is estimated that the SLAM system cannot work and cannot be restarted.
  • the first aspect of the embodiment of the present application provides a pose prediction method, the method comprising:
  • the movement speed parameter includes the angular velocity and linear acceleration of the target device, and the previous moment is a historical moment before the target device loses visual information;
  • the second aspect of the embodiment of the present application provides a device for pose prediction, the device comprising:
  • a parameter acquisition module which acquires a movement speed parameter of the target device at a previous moment; wherein, the movement speed parameter includes the angular velocity and linear acceleration of the target device, and the previous moment is a historical moment before the target device loses visual information;
  • An increment determination module configured to estimate the predicted displacement increment of the target device at each current moment by using the motion speed parameter and the preset displacement prediction model; wherein, the current moment is when the target device loses visual information The latest moment and every moment after the loss of visual information;
  • a pose and trajectory prediction module configured to calculate a predicted pose of the target device corresponding to each current moment according to the predicted displacement increment of the target device at each current moment, and construct a predicted motion trajectory of the target device;
  • An optimization module configured to optimize the predicted pose and the predicted motion trajectory of the target device, and acquire target poses and corresponding target motion trajectories of the target device at each current moment.
  • the third aspect of the embodiments of the present application provides a pose prediction system, including an inertial navigation sensor and a terminal device, wherein the inertial navigation sensor is used for the movement speed parameter of the terminal device at the previous moment, and the The terminal device is used to realize the steps of the above method.
  • a fourth aspect of the embodiments of the present application provides a terminal device, including a memory, a processor, and a computer program stored in the memory and operable on the processor, and the processor implements the steps of the above method when executing the computer program.
  • a fifth aspect of the embodiments of the present application provides a computer-readable storage medium, where a computer program is stored in the computer-readable storage medium, and when the computer program is executed by a processor, the steps of the above method are implemented.
  • the present application has the beneficial effect that the target device’s pose can be estimated only by using the target device’s motion velocity parameters without the need for visual information of the target device; In the case of the visual information of the target device, the pose of the target device can still be estimated, so as to ensure that the SLAM system can continue to work using the estimated pose, thereby ensuring the robustness of the SLAM system.
  • Fig. 1 is a schematic diagram of a pose prediction system provided by an embodiment of the present application
  • Fig. 2 is a flow chart of the pose prediction method provided by the embodiment of the present application.
  • Fig. 3 is a schematic diagram of the network architecture of the displacement prediction model provided by the embodiment of the present application.
  • Fig. 4 is a schematic diagram of an optimized pose graph provided by an embodiment of the present application.
  • FIG. 5 is a block diagram of a pose prediction device provided in an embodiment of the present application.
  • Fig. 6 is a schematic diagram of a computer device provided by an embodiment of the present application.
  • the present invention provides a pose prediction method, which obtains the motion speed parameters of the target device at the previous moment; uses the motion speed parameters and the preset displacement prediction model to estimate the predicted displacement increment of the target device at each current moment;
  • the predicted displacement increment calculates the predicted pose of the target device at each current moment, and constructs the predicted motion trajectory; optimizes the predicted pose and predicted motion trajectory at each current moment, and obtains the target pose and the corresponding target device at each current moment. target trajectory.
  • Fig. 1 is a schematic diagram of a pose prediction system provided in this embodiment, which includes an inertial sensor (Inertial Measurement Unit, IMU) 1 and a terminal device 2, wherein, in one implementation, the terminal device 2 can be a server, or Various electronic devices that support data processing, including but not limited to smartphones, sweepers, tablets, laptops and desktop computers, etc.
  • the IMU1 can be set on a target device, and the target device can be a device that needs to be positioned, for example, the target device can be an AR/VR device, a robot, a vehicle, and the like.
  • the IMU1 can measure the movement speed parameter of the target device at the previous moment, and send the movement speed parameter to the terminal device 2 .
  • the terminal device 2 After the terminal device 2 obtains the movement speed parameter of the target device at the previous moment, it can obtain the predicted displacement increment of the target device according to the movement speed parameter and the preset displacement prediction model; the terminal device 2 can obtain the predicted displacement increment of the target device , to determine the predicted pose and predicted motion trajectory of the target device; the terminal device 2 can optimize the predicted pose and predicted motion trajectory of the target device to obtain the target pose and target motion trajectory of the target device.
  • Fig. 2 is a flow chart of a pose prediction method provided by an embodiment of the present application.
  • a pose prediction method in FIG. 2 may be executed by the data processing device 2 in FIG. 1 .
  • the pose prediction method includes:
  • S201 Obtain a movement speed parameter of the target device at a previous moment; the movement speed parameter includes the angular velocity and the linear acceleration of the target device.
  • the target device can be understood as a device that needs to predict a pose, for example, the target device can be an AR/VR device, a mobile phone, an autonomously moving robot, or an unmanned vehicle.
  • the target device may be equipped with a SLAM (Simultaneous Localization and Mapping) system, wherein the SLAM system includes an image acquisition device (such as a camera) and an IMU.
  • SLAM Simultaneous Localization and Mapping
  • the image acquisition device may collect the visual information of the target device every preset time period, for example, the image acquisition device may collect the visual information of the target device every 30 ms.
  • the visual information of the target device may be understood as image frames around the target device, for example, image frames of the front side, rear side, left side, and right side of the target device.
  • the IMU can measure the motion speed parameters of the target device at preset intervals, for example, the IMU can measure the motion speed parameters of the target device at a frequency of 100 Hz.
  • the moving speed parameter of the target device can be understood as the speed data of the target device during the moving process.
  • the movement speed parameter may include the angular velocity and linear acceleration in the device coordinate system of the target device, that is, the angular velocity and linear acceleration in the body coordinate system (ie, body frame) of the target device.
  • the accelerometers can be used to detect the respective linear accelerations of the target device on the x-axis, y-axis, and z-axis in the device coordinate system.
  • the gyroscope detects the respective angular velocities of the target device on the x-axis, y-axis, and z-axis in the device coordinate system; it can be understood that the movement speed parameters can include the x-axis, y-axis, and z-axis in the device coordinate system of the target device respective angular velocity and linear acceleration.
  • the IMU may also be a 9-axis IMU, that is, include a magnetometer, and the IMU is not specifically limited in this embodiment.
  • this embodiment uses the IMU to obtain the motion velocity parameters of the target device at the previous moment to predict the pose of the target device at the current moment. It should be noted that the previous moment is a historical moment before the target device loses the visual information.
  • the method may further include the following steps:
  • the acquisition of the movement speed parameter of the target device at the previous moment is performed.
  • the visual information of the target device if the visual information of the target device is not detected, it means that the visual information of the target device is lost, so it is not possible to use the visual information to estimate the pose of the target device; at this time, it is necessary to obtain the target device The motion speed parameters at the previous moment, so that the motion speed parameters of the target device can be used to estimate the pose of the target device.
  • the target device is equipped with a SLAM system
  • the SLAM system in the main function of the SLAM tracking thread, when the system state of the SLAM system is normal, the SLAM system is in the normal tracking state; when the visual information is lost, the SLAM system enters the tracking loss state , at this time, the motion velocity parameters collected by the IMU can be used to perform integral operations to obtain the recursive pose, but since the reliability of the recursive position obtained through the integral operation is not high, when the loss of visual information exceeds the preset duration ( For example, after 1s), if the visual information has not returned to normal, the SLAM system enters the tracking loss state. At this time, the step of obtaining the motion speed parameter of the target device at the previous moment needs to be executed.
  • S202 Estimate the predicted displacement increment of the target device at each current moment by using the motion speed parameter and the preset displacement prediction model.
  • the current moment is the latest moment when the target device loses the visual information and every moment after the visual information is lost.
  • the preset displacement prediction model can be used to determine the subsequent predicted displacement increment of the target device at each current moment according to the movement speed parameters, wherein the target device
  • the predicted displacement increment at the current moment can be understood as the predicted variation of displacement at each moment during the time when the target device loses visual information.
  • the target rotation matrix may be determined according to the angular velocity of the target device in the device coordinate system at the previous moment.
  • the angular velocity under the device coordinate system of the target device is integrated to obtain the target rotation matrix, wherein the target rotation matrix can be understood as converting the motion velocity parameter from the device coordinate system (ie body frame) to the preset world of the SLAM system
  • the rotation matrix that is, the rotation Rwb) in the coordinate system (world frame).
  • target rotation matrix to rotate the angular velocity and linear acceleration in the equipment coordinate system to obtain the angular velocity and linear acceleration in the world coordinate system, and input the angular velocity and linear acceleration in the world coordinate system into the preset displacement prediction model to obtain the target equipment Predicted displacement increments at each current moment.
  • the input of the preset displacement prediction model can be the angular velocity and linear acceleration of the x-axis, y-axis, and z-axis in the world coordinate system of a fixed time length (such as 1s); it should be noted that, assuming a fixed time length is 1s, since the data of angular velocity and linear acceleration at one moment is 6-dimensional, if the acquisition frequency of angular velocity and linear acceleration is 250HZ, then there are 250 pieces of angular velocity and linear acceleration data collected in 1s. It is understandable that the preset The input tensor size of the displacement prediction model is batch_size ⁇ 6 ⁇ 250.
  • the output of the preset displacement prediction model is the predicted displacement increment of the x-axis, y-axis, and z-axis of the target device in the world coordinate system, that is, within a period of time, the x-axis and y-axis of the target device in the world coordinate system , The displacement increment on the z axis.
  • the preset displacement prediction model includes a plurality of cascaded convolutional layers and an output layer, as shown in FIG.
  • the layers are connected, and the output layer is composed of a global average pooling layer (AngPool1d) and a convolutional layer (Conv1d) in series, thereby retaining the spatial structure of the network, greatly reducing the number of parameters of the model, reducing the reasoning time, and preventing The technical effect of overfitting.
  • the displacement prediction model also includes a BN layer and a relu layer, but they are omitted in Figure 3, where c in Figure 3 represents the number of channels, and k represents the size of the volume.
  • the model can be supervised by using a loss function, wherein the loss function is preferably a mean-square error (Mean-Square Error, MSE) loss function.
  • MSE LOSS that is, the MSE loss function
  • L mse is MSE LOSS
  • i ⁇ [0,n] representing the i-th moment.
  • the network structure of the displacement prediction model can also be a one-dimensional form of resnet18, TCN, LSTM and other neural network structures, which are not limited here.
  • S203 Calculate the predicted pose of the target device corresponding to each current moment according to the predicted displacement increment of the target device at each current moment, and construct a predicted motion trajectory of the target device.
  • the predicted displacement increment at each current moment and the pose information of the target device at the previous moment can be used to obtain the prediction of the target device at each current moment Pose information, and then use all the predicted pose information at the current moment to construct the predicted motion trajectory of the target device.
  • the predicted displacement increment of the target device includes a predicted displacement increment on the X axis, a predicted displacement increment on the Y axis, and a predicted displacement increment on the Z axis.
  • the target is determined according to the predicted displacement increment on the X-axis, the predicted displacement increment on the Y-axis, the predicted displacement increment on the Z-axis, and the pose information of the target device at the previous moment.
  • the predicted pose of the device on the X-axis, the predicted pose on the Y-axis, and the predicted pose on the Z-axis at the current moment for example, the pose information of the target device at the previous moment (that is, the x-axis, y-axis, Coordinates on the z-axis) add the predicted displacement increment of the target device on the X-axis, the predicted displacement increment on the Y-axis, and the predicted displacement increment on the Z-axis to obtain the predicted position of the target device on the X-axis at the current moment pose, the predicted pose on the Y axis, and the predicted pose on the Z axis (that is, the predicted coordinate value).
  • the predicted position of the target device at each current moment after losing visual information can be determined
  • Attitude information that is, the predicted position of the target device on the X-axis, the predicted position on the Y-axis, and the predicted position on the Z-axis at the current moment can be used as the predicted pose information of the target device at the current moment; the predicted position of the target device at the current moment
  • the pose information is regarded as the pose information of the previous moment, and the above calculation is repeated to predict the predicted pose of the next current moment, so as to obtain the predicted pose of each current moment during the time period when the target device loses visual information.
  • the predicted motion trajectory of the target device is constructed using the predicted pose information of the displacement at each current moment, as shown by the solid line in FIG. 4 .
  • this application can also perform pose estimation through zero-speed update combined with Kalman filter (EKF), Step Counting and other methods, and there is no limitation here.
  • EKF Kalman filter
  • the target device's motion velocity parameters can be used to estimate the predicted pose of the target device, in order to ensure the accuracy of the determined running trajectory of the target device, it is necessary to continuously check whether there is a visual image of the target device collected. Information, so that the predicted motion trajectory of the target device can be corrected according to the collected visual information of the target device.
  • step S204 also includes:
  • the predicted pose of the target device is optimized according to the image frame in the visual information to obtain the target pose.
  • the relocation condition may be that the target device can be positioned according to the visual information
  • determine the target pose of the target device that is, after detecting the visual information, judge whether the image frame in the visual information matches successfully with an image frame in the global map or local map pre-established by the SLAM system; if the visual information The image frame in the SLAM system is successfully matched with an image frame in the global map or local map pre-established by the SLAM system, indicating that the visual information satisfies the relocation condition, so that the target device can be located according to the visual information and the target of the target device can be determined
  • the image frame in the visual information cannot match any image frame in the pre-established global map or local map of the SLAM system, it means that the visual information does not meet the relocation conditions, so it cannot be based on
  • the visual information locates the target device to determine the target pose of
  • a pose optimization may be performed on the predicted pose of the target device according to the image frames in the visual information to obtain a target pose.
  • the target pose corresponding to the target device may be determined first according to the image frame in the visual information.
  • the image frame that matches the image frame in the visual information can be determined in the global map or local map pre-established in the SLAM system, wherein the similarity between the image frame in the visual information and the matched image frame is greater than Preset threshold, or, the matching image frame is the largest similarity between all image frames in the map and the image frame in the visual information; then, according to the image frame in the visual information and the matching image frame in the map Image frames, determine the ideal pose of the target device.
  • the predicted pose and predicted motion trajectory before the visual information is not detected are corrected and optimized according to the ideal pose, and the target pose and target motion trajectory of the target device are obtained.
  • the predicted pose before no visual information is detected can be obtained first, that is, the predicted pose of the target device is determined using a preset displacement prediction model; and then the predicted pose before no visual information is detected according to the ideal pose
  • the predicted motion trajectory corresponding to the pose and predicted pose is optimized to obtain the target pose and target motion trajectory of the target device.
  • the predicted pose is used as the optimization variable
  • the error of the relative motion estimation between the predicted pose and the ideal pose that is, the predicted motion trajectory corresponding to the predicted pose of the target device
  • the The optimization model is shown in the following formula:
  • T i is the predicted pose of the target device at the i-th moment
  • is the covariance of the residual
  • min() represents minimization
  • ⁇ T m is the predicted pose The observed value of the relative pose with the ideal pose
  • r() is a residual function, which represents the residual between the relative pose between the predicted pose to be optimized and the ideal pose and its observed value, specifically: Then, the optimization model is solved to obtain the target motion trajectory, where the starting point and end point of the visual information loss stage in the target motion trajectory are the detected pose before the loss of visual information and the ideal pose of the target device, respectively.
  • the error of the predicted motion trajectory corresponding to the predicted pose of the target device will be reduced, and the head and tail of the trajectory in the visual information loss stage in the optimized target motion trajectory will be compared with the SLAM system based on the collected visual information.
  • Normal estimated trajectory alignment for example, as shown in Figure 4, the solid line segment in the visual information loss stage is the predicted motion trajectory corresponding to the predicted pose, and the dashed line segment is the optimized target motion trajectory.
  • the visual information of the target device is detected, and the visual information does not satisfy the relocation condition, then continue to perform the acquisition of the movement speed parameter of the target device at the previous moment until the visual information of the target device is detected, and The visual information satisfies the relocation criteria.
  • a preset time period such as 20s
  • Fig. 5 is a schematic diagram of a pose prediction device provided by an embodiment of the present application. As shown in Figure 5, the pose prediction device includes:
  • the parameter acquisition module 501 is configured to acquire the movement speed parameter of the target device at a previous moment; wherein, the movement speed parameter includes the angular velocity and linear acceleration of the target device, and the previous moment is a historical moment before the target device loses visual information;
  • the incremental calculation module 502 is used to estimate the predicted displacement increment of the target device at each current moment by using the motion speed parameter and the preset displacement prediction model; wherein, the current moment is the latest moment when the target device loses visual information and after the visual information is lost every moment of
  • the pose and trajectory prediction module 503 is used to calculate the predicted pose of the target device corresponding to each current moment according to the predicted displacement increment of each current moment of the target device, and construct the predicted motion trajectory of the target device;
  • the optimization module 504 is configured to optimize the predicted pose and predicted motion trajectory of the target device, and obtain the target pose and corresponding target motion trajectory of the target device at each current moment.
  • the incremental calculation module 502 is specifically used for:
  • the predicted displacement increment of the target device includes a predicted displacement increment on the X axis, a predicted displacement increment on the Y axis, and a predicted displacement increment on the Z axis;
  • the position determination module 503 is specifically used for:
  • the predicted displacement increment on the X-axis determines the target device's position The predicted position on the X axis, the predicted position on the Y axis, and the predicted position on the Z axis;
  • the predicted positions of the target device on the X-axis the predicted position on the Y-axis and the predicted position on the Z-axis, the predicted poses of the target device at each current moment are determined.
  • the device also includes a detection module; the detection module is used for:
  • optimization module specifically for:
  • the predicted pose and predicted motion trajectory of the target device are optimized according to the image frames in the visual information to obtain the target pose and target motion trajectory.
  • optimization module specifically for:
  • the predicted pose of the target device is optimized to obtain the target pose.
  • optimization module is also used to:
  • optimization module is also used to:
  • FIG. 6 is a schematic diagram of a terminal device 6 provided by an embodiment of the present application.
  • the terminal device 6 includes: a processor 601 , a memory 602 and a computer program 603 stored in the memory 602 and capable of running on the processor 601 .
  • the processor 601 executes the computer program 603
  • the steps in the foregoing method embodiments are implemented.
  • the processor 601 executes the computer program 603 the functions of the modules/modules in the foregoing device embodiments are realized.
  • the computer program 603 can be divided into one or more modules/modules, and one or more modules/modules are stored in the memory 602 and executed by the processor 601 to complete the present application.
  • One or more modules/modules may be a series of computer program instruction segments capable of performing specific functions, and the instruction segments are used to describe the execution process of the computer program 603 in the computer device 6.
  • the terminal device 6 may include, but not limited to, a processor 601 and a memory 602 .
  • FIG. 6 is only an example of the terminal device 6, and does not constitute a limitation on the terminal device 6. It may include more or less components than those shown in the figure, or combine certain components, or different components.
  • computer equipment may also include input and output equipment, network access equipment, bus, and so on.
  • the processor 601 can be a central processing unit (Central Processing Unit, CPU), or other general processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), on-site Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
  • CPU Central Processing Unit
  • DSP Digital Signal Processor
  • ASIC Application Specific Integrated Circuit
  • FPGA Field-Programmable Gate Array
  • a general-purpose processor may be a microprocessor, or the processor may be any conventional processor, or the like.
  • the storage 602 may be an internal storage module of the terminal device 6 , for example, a hard disk or memory of the terminal device 6 .
  • the memory 602 can also be an external storage device of the computer device 6, for example, a plug-in hard disk equipped on the terminal device 6, a smart memory card (Smart Media Card, SMC), a secure digital (Secure Digital, SD) card, a flash memory card ( Flash Card), etc. Further, the memory 602 may also include both an internal storage module of the terminal device 6 and an external storage device.
  • the memory 602 is used to store computer programs and other programs and data required by the computer equipment.
  • the memory 602 can also be used to temporarily store data that has been output or will be output.
  • modules and algorithm steps of the examples described in conjunction with the embodiments disclosed herein can be implemented by electronic hardware, or a combination of computer software and electronic hardware. Whether these functions are executed by hardware or software depends on the specific application and design constraints of the technical solution. Skilled artisans may use different methods to implement the described functions for each specific application, but such implementation should not be regarded as exceeding the scope of the present application.
  • the disclosed apparatus/computer equipment and methods can be implemented in other ways.
  • the device/computer device embodiments described above are only illustrative, for example, the division of modules or modules is only a logical function division, and there may be other division methods in actual implementation, and multiple modules or components can be Incorporation may either be integrated into another system, or some features may be omitted, or not implemented.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be through some interfaces, and the indirect coupling or communication connection of devices or modules may be in electrical, mechanical or other forms.
  • a module described as a separate component may or may not be physically separated, and a component shown as a module may or may not be a physical module, that is, it may be located in one place, or may also be distributed to multiple network modules. Part or all of the modules can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
  • each functional module in each embodiment of the present application may be integrated into one processing module, each module may exist separately physically, or two or more modules may be integrated into one module.
  • the above-mentioned integrated modules can be implemented in the form of hardware or in the form of software function modules.
  • the integrated modules/modules are implemented in the form of software function modules and sold or used as independent products, they can be stored in a computer-readable storage medium. Based on this understanding, the present application realizes all or part of the processes in the methods of the above embodiments, and can also be completed by instructing related hardware through computer programs.
  • the computer programs can be stored in computer-readable storage media, and the computer programs can be processed. When executed by the controller, the steps in the above-mentioned method embodiments can be realized.
  • a computer program may include computer program code, which may be in source code form, object code form, executable file, or some intermediate form or the like.
  • the computer-readable medium may include: any entity or device capable of carrying computer program code, recording medium, U disk, removable hard disk, magnetic disk, optical disk, computer memory, read-only memory (Read-Only Memory, ROM), random access Memory (Random Access Memory, RAM), electrical carrier signal, telecommunication signal and software distribution medium, etc. It should be noted that the content contained in computer readable media may be appropriately increased or decreased according to the requirements of legislation and patent practice in the jurisdiction. For example, in some jurisdictions, computer readable media may not Including electrical carrier signals and telecommunication signals.

Abstract

A pose prediction method and apparatus. Under the condition that a target device loses visual information, the pose of the target device is estimated only using a motion speed parameter of the target device. Even if the visual information of the target device is not collected, the pose of the target device is still estimated, such that a SLAM system can be ensured to continue working and running by utilizing the estimated pose, thereby ensuring the robustness of the SLAM system.

Description

一种位姿预测方法及装置A pose prediction method and device
本申请要求于2022年1月26日提交中国专利局,申请号为202210093752.1,发明名称为“一种位姿预测方法及装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application with the application number 202210093752.1 and the title of the invention "a method and device for pose prediction" submitted to the China Patent Office on January 26, 2022, the entire contents of which are incorporated in this application by reference middle.
技术领域technical field
本申请涉及定位技术领域,尤其涉及一种位姿预测方法及装置。The present application relates to the technical field of positioning, and in particular to a pose prediction method and device.
背景技术Background technique
随着科技发展,AR/VR、机器人、无人驾驶等前沿科技迅猛发展,而AR/VR、机器人、无人驾驶技术领域中均涉及到自主定位技术。自主定位技术用于确定设备(机器人/无人车/手机等)自身的位姿。With the development of science and technology, cutting-edge technologies such as AR/VR, robots, and unmanned driving are developing rapidly, and the fields of AR/VR, robots, and unmanned driving technology all involve autonomous positioning technology. Autonomous positioning technology is used to determine the pose of the device (robot/unmanned vehicle/mobile phone, etc.).
目前,在室外定位技术中,通常采用GPS定位技术,而在室内或者GPS信号不好的地方,需要用其他定位技术,比如室内的UWB、蓝牙、动作捕捉系统等等。如今,SLAM(即时定位与地图构建,Simultaneous Localization and Mapping)逐渐成为一种室内重要的定位技术,SLAM是一种利用自身搭载的传感器感知环境,实时计算自身位姿,并构建增量式地图的技术,其不需要改造外部环境,定位精度可达厘米级别,其应用领域可以包括AR/VR、机器人、无人驾驶、无人机等。At present, in outdoor positioning technology, GPS positioning technology is usually used, while indoors or places with poor GPS signals need to use other positioning technologies, such as indoor UWB, Bluetooth, motion capture systems, etc. Today, SLAM (Simultaneous Localization and Mapping) has gradually become an important indoor positioning technology. SLAM is a method that uses its own sensors to perceive the environment, calculate its own pose in real time, and build incremental maps. Technology, it does not need to modify the external environment, and its positioning accuracy can reach centimeter level. Its application fields can include AR/VR, robots, unmanned driving, drones, etc.
现有的SLAM系统的定位非常依赖于视觉信息,如果某段时间内出现环境纹理弱、设备运动速度较快等情况,会出现SLAM系统所采集到的视觉信息中的图像为弱纹理、遮挡或者模糊,从而影响SLAM系统的正常运行,例如,当未采集到视觉信息时,SLAM系统不能进行设备的位姿估计,从而导致SLAM 系统无法工作运行以及无法重启,而且必须等到重新采集到视觉信息之后才能重启并工作运行。因此,亟需一种能够解决由于未采集到视觉信息时,SLAM系统不能进行设备的位姿估计,而导致SLAM系统无法工作运行以及无法重启的问题的技术方案。The positioning of the existing SLAM system is very dependent on visual information. If the environment texture is weak and the equipment moves faster within a certain period of time, the image in the visual information collected by the SLAM system will appear as weak texture, occlusion or Blurred, which affects the normal operation of the SLAM system. For example, when the visual information is not collected, the SLAM system cannot estimate the pose of the device, which makes the SLAM system unable to work and cannot be restarted, and must wait until the visual information is re-collected. to restart and work. Therefore, there is an urgent need for a technical solution that can solve the problem that the SLAM system cannot work and cannot be restarted because the SLAM system cannot estimate the pose of the device when the visual information is not collected.
发明内容Contents of the invention
有鉴于此,本申请实施例提供了一种位姿预测方法、装置、计算机设备及计算机可读存储介质,以解决现有技术中由于未采集到视觉信息时,SLAM系统不能进行设备的位姿估计,而导致SLAM系统无法工作运行以及无法重启的问题。In view of this, the embodiment of the present application provides a pose prediction method, device, computer equipment, and computer-readable storage medium to solve the problem that the SLAM system cannot perform the pose prediction of the device when no visual information is collected in the prior art. It is estimated that the SLAM system cannot work and cannot be restarted.
本申请实施例的第一方面,提供了一种位姿预测方法,所述方法包括:The first aspect of the embodiment of the present application provides a pose prediction method, the method comprising:
获取目标设备前一时刻的运动速度参数;其中,所述运动速度参数包括目标设备的角速度和线加速度,所述前一时刻为所述目标设备丢失视觉信息前的一历史时刻;Obtaining the movement speed parameter of the target device at a previous moment; wherein the movement speed parameter includes the angular velocity and linear acceleration of the target device, and the previous moment is a historical moment before the target device loses visual information;
利用所述运动速度参数和预设位移预测模型,估计所述目标设备各当前时刻的预测位移增量;其中,所述当前时刻为所述目标设备丢失视觉信息时的最新时刻及丢失视觉信息后的每个时刻;Estimate the predicted displacement increment of the target device at each current moment by using the motion speed parameter and the preset displacement prediction model; wherein, the current moment is the latest moment when the target device loses visual information and after losing visual information every moment of
根据所述目标设备各当前时刻的所述预测位移增量计算所述目标设备各当前时刻对应的预测位姿,并构建所述目标设备的预测运动轨迹;calculating the predicted pose corresponding to each current moment of the target device according to the predicted displacement increment of the target device at each current moment, and constructing a predicted motion trajectory of the target device;
对所述目标设备的所述预测位姿及所述预测运动轨迹进行优化,获取所述目标设备各当前时刻的目标位姿及对应的目标运动轨迹。Optimizing the predicted pose and the predicted motion trajectory of the target device, and acquiring the target pose and corresponding target motion trajectory of the target device at each current moment.
本申请实施例的第二方面,提供了一种位姿预测装置,所述装置包括:The second aspect of the embodiment of the present application provides a device for pose prediction, the device comprising:
参数获取模块,获取目标设备前一时刻的运动速度参数;其中,所述运动速度参数包括目标设备的角速度和线加速度,所述前一时刻为所述目标设备丢失视觉信息前的一历史时刻;A parameter acquisition module, which acquires a movement speed parameter of the target device at a previous moment; wherein, the movement speed parameter includes the angular velocity and linear acceleration of the target device, and the previous moment is a historical moment before the target device loses visual information;
增量确定模块,用于利用所述运动速度参数和预设位移预测模型,估计所述目标设备各当前时刻的预测位移增量;其中,所述当前时刻为所述目标设备 丢失视觉信息时的最新时刻及丢失视觉信息后的每个时刻;An increment determination module, configured to estimate the predicted displacement increment of the target device at each current moment by using the motion speed parameter and the preset displacement prediction model; wherein, the current moment is when the target device loses visual information The latest moment and every moment after the loss of visual information;
位姿与轨迹预测模块,用于根据所述目标设备各当前时刻的所述预测位移增量计算所述目标设备各当前时刻对应的预测位姿,并构建所述目标设备的预测运动轨迹;A pose and trajectory prediction module, configured to calculate a predicted pose of the target device corresponding to each current moment according to the predicted displacement increment of the target device at each current moment, and construct a predicted motion trajectory of the target device;
优化模块,用于对所述目标设备的所述预测位姿及所述预测运动轨迹进行优化,获取所述目标设备各当前时刻的目标位姿及对应的目标运动轨迹。An optimization module, configured to optimize the predicted pose and the predicted motion trajectory of the target device, and acquire target poses and corresponding target motion trajectories of the target device at each current moment.
本申请实施例的第三方面,提供了一种位姿预测系统,包括惯导传感器及终端设备,其中,所述惯导传感器用于所述终端设备在前一时刻的运动速度参数,所述终端设备用于实现上述方法的步骤。The third aspect of the embodiments of the present application provides a pose prediction system, including an inertial navigation sensor and a terminal device, wherein the inertial navigation sensor is used for the movement speed parameter of the terminal device at the previous moment, and the The terminal device is used to realize the steps of the above method.
本申请实施例的第四方面,提供了一种终端设备,包括存储器、处理器以及存储在存储器中并且可以在处理器上运行的计算机程序,该处理器执行计算机程序时实现上述方法的步骤。A fourth aspect of the embodiments of the present application provides a terminal device, including a memory, a processor, and a computer program stored in the memory and operable on the processor, and the processor implements the steps of the above method when executing the computer program.
本申请实施例的第五方面,提供了一种计算机可读存储介质,该计算机可读存储介质存储有计算机程序,该计算机程序被处理器执行时实现上述方法的步骤。A fifth aspect of the embodiments of the present application provides a computer-readable storage medium, where a computer program is stored in the computer-readable storage medium, and when the computer program is executed by a processor, the steps of the above method are implemented.
本申请与现有技术相比存在的有益效果是:实现不需要目标设备的视觉信息,而仅利用目标设备的运动速度参数,便可以估计目标设备的位姿;这样,可以实现在未采集到目标设备的视觉信息的情况下,依然可以估计目标设备的位姿,从而保证SLAM系统可以利用所估计的位姿继续工作运行,进而保证了SLAM系统的鲁棒性。Compared with the prior art, the present application has the beneficial effect that the target device’s pose can be estimated only by using the target device’s motion velocity parameters without the need for visual information of the target device; In the case of the visual information of the target device, the pose of the target device can still be estimated, so as to ensure that the SLAM system can continue to work using the estimated pose, thereby ensuring the robustness of the SLAM system.
附图说明Description of drawings
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其它的附图。In order to more clearly illustrate the technical solutions in the embodiments of the present application, the accompanying drawings that need to be used in the descriptions of the embodiments or the prior art will be briefly introduced below. Obviously, the accompanying drawings in the following description are only for the present application For some embodiments, those skilled in the art can also obtain other drawings based on these drawings without creative efforts.
图1是本申请实施例提供的位姿预测系统示意图;Fig. 1 is a schematic diagram of a pose prediction system provided by an embodiment of the present application;
图2是本申请实施例提供的位姿预测方法的流程图;Fig. 2 is a flow chart of the pose prediction method provided by the embodiment of the present application;
图3是本申请实施例提供的位移预测模型的网络架构示意图;Fig. 3 is a schematic diagram of the network architecture of the displacement prediction model provided by the embodiment of the present application;
图4是本申请实施例提供的优化位姿图的示意图;Fig. 4 is a schematic diagram of an optimized pose graph provided by an embodiment of the present application;
图5是本申请实施例提供的位姿预测装置的框图;FIG. 5 is a block diagram of a pose prediction device provided in an embodiment of the present application;
图6是本申请实施例提供的计算机设备的示意图。Fig. 6 is a schematic diagram of a computer device provided by an embodiment of the present application.
具体实施方式Detailed ways
以下描述中,为了说明而不是为了限定,提出了诸如特定系统结构、技术之类的具体细节,以便透彻理解本申请实施例。然而,本领域的技术人员应当清楚,在没有这些具体细节的其它实施例中也可以实现本申请。在其它情况中,省略对众所周知的系统、装置、电路以及方法的详细说明,以免不必要的细节妨碍本申请的描述。In the following description, specific details such as specific system structures and technologies are presented for the purpose of illustration rather than limitation, so as to thoroughly understand the embodiments of the present application. It will be apparent, however, to one skilled in the art that the present application may be practiced in other embodiments without these specific details. In other instances, detailed descriptions of well-known systems, devices, circuits, and methods are omitted so as not to obscure the description of the present application with unnecessary detail.
本发明提供了一种位姿预测方法,获取目标设备前一时刻的运动速度参数;利用运动速度参数和预设位移预测模型,估计目标设备各当前时刻的预测位移增量;根据各当前时刻的预测位移增量计算目标设备各当前时刻对应的预测位姿,并构建预测运动轨迹;对各当前时刻的预测位姿及预测运动轨迹进行优化,获取目标设备各当前时刻的目标位姿及对应的目标运动轨迹。The present invention provides a pose prediction method, which obtains the motion speed parameters of the target device at the previous moment; uses the motion speed parameters and the preset displacement prediction model to estimate the predicted displacement increment of the target device at each current moment; The predicted displacement increment calculates the predicted pose of the target device at each current moment, and constructs the predicted motion trajectory; optimizes the predicted pose and predicted motion trajectory at each current moment, and obtains the target pose and the corresponding target device at each current moment. target trajectory.
图1为本实施例提供的位姿预测系统示意图,该系统中包括惯性传感器(Inertial Measurement Unit,IMU)1和终端设备2,其中,在一种实现方式中,终端设备2可以为服务器,或者支持数据处理的各种电子设备,包括但不限于智能手机、扫地机、平板电脑、膝上型便携计算机和台式计算机等。IMU1可以设置在目标设备上,目标设备可以为需要进行定位的设备,例如目标设备可以为AR/VR设备、机器人、车辆等。Fig. 1 is a schematic diagram of a pose prediction system provided in this embodiment, which includes an inertial sensor (Inertial Measurement Unit, IMU) 1 and a terminal device 2, wherein, in one implementation, the terminal device 2 can be a server, or Various electronic devices that support data processing, including but not limited to smartphones, sweepers, tablets, laptops and desktop computers, etc. The IMU1 can be set on a target device, and the target device can be a device that needs to be positioned, for example, the target device can be an AR/VR device, a robot, a vehicle, and the like.
具体地,IMU1可以测量目标设备在前一时刻的运动速度参数,并将该运动速度参数向终端设备2发送。终端设备2获取该目标设备在前一时刻的运动速度参数后,可以根据运动速度参数和预设位移预测模型,得到目标设备的预 测位移增量;终端设备2可以根据目标设备的预测位移增量,确定目标设备的预测位姿及预测运动轨迹;终端设备2可以对目标设备的预测位姿及预测运动轨迹进行优化进而获取目标设备的目标位姿及目标运动轨迹。Specifically, the IMU1 can measure the movement speed parameter of the target device at the previous moment, and send the movement speed parameter to the terminal device 2 . After the terminal device 2 obtains the movement speed parameter of the target device at the previous moment, it can obtain the predicted displacement increment of the target device according to the movement speed parameter and the preset displacement prediction model; the terminal device 2 can obtain the predicted displacement increment of the target device , to determine the predicted pose and predicted motion trajectory of the target device; the terminal device 2 can optimize the predicted pose and predicted motion trajectory of the target device to obtain the target pose and target motion trajectory of the target device.
需要注意的是,上述应用场景仅是为了便于理解本申请而示出,本申请的实施方式在此方面不受任何限制。相反,本申请的实施方式可以应用于适用的任何场景。It should be noted that the above application scenarios are only illustrated for the convenience of understanding the present application, and the implementation manners of the present application are not limited in this regard. On the contrary, the embodiments of the present application can be applied to any applicable scene.
图2是本申请实施例提供的一种位姿预测方法的流程图。图2的一种位姿预测方法可以由图1的数据处理设备2执行。如图2所示,该位姿预测方法包括:Fig. 2 is a flow chart of a pose prediction method provided by an embodiment of the present application. A pose prediction method in FIG. 2 may be executed by the data processing device 2 in FIG. 1 . As shown in Figure 2, the pose prediction method includes:
S201:获取目标设备在前一时刻的运动速度参数;运动速度参数包括目标设备的角速度和线加速度。S201: Obtain a movement speed parameter of the target device at a previous moment; the movement speed parameter includes the angular velocity and the linear acceleration of the target device.
在本实施例中,目标设备可以理解为需要预测位姿的设备,例如,目标设备可以为AR/VR设备、手机、自主移动的机器人、无人驾驶车辆。在一种实现方式中,目标设备中可以搭载有SLAM(Simultaneous Localization and Mapping)系统,其中,SLAM系统包括图像采集装置(比如摄像头)和IMU。In this embodiment, the target device can be understood as a device that needs to predict a pose, for example, the target device can be an AR/VR device, a mobile phone, an autonomously moving robot, or an unmanned vehicle. In an implementation manner, the target device may be equipped with a SLAM (Simultaneous Localization and Mapping) system, wherein the SLAM system includes an image acquisition device (such as a camera) and an IMU.
具体地,图像采集装置可以每隔预设时长采集目标设备的视觉信息,比如,图像采集装置可以每隔30ms采集目标设备的视觉信息。其中,目标设备的视觉信息可以理解为目标设备的周围的图像帧,比如,目标设备的前侧、后侧、左侧、右侧的图像帧。Specifically, the image acquisition device may collect the visual information of the target device every preset time period, for example, the image acquisition device may collect the visual information of the target device every 30 ms. Wherein, the visual information of the target device may be understood as image frames around the target device, for example, image frames of the front side, rear side, left side, and right side of the target device.
IMU可以每隔预设时长测量目标设备的运动速度参数,比如IMU可以以100Hz的频率测量目标设备的运动速度参数。其中,目标设备的运动速度参数可以理解为目标设备在运动过程的速度数据。在一种实现方式中,运动速度参数可以包括目标设备的设备坐标系下的角速度和线加速度,即在目标设备的本体坐标系(即body frame)下的角速度和线加速度。由于IMU包括了三个单轴的加速度计和三个单轴的陀螺仪,因此,可以利用加速度计检测目标设备在设备坐标系中x轴、y轴、z轴上各自的线加速度,可以利用陀螺仪检测目标设备 在设备坐标系中x轴、y轴、z轴上各自的角速度;可以理解的是,运动速度参数可以包括目标设备的设备坐标系下的x轴、y轴、z轴上各自的角速度和线加速度。需要说明的是,IMU也可以为9轴的IMU,即包括磁力计,在本实施例中不对IMU进行具体限定。The IMU can measure the motion speed parameters of the target device at preset intervals, for example, the IMU can measure the motion speed parameters of the target device at a frequency of 100 Hz. Wherein, the moving speed parameter of the target device can be understood as the speed data of the target device during the moving process. In an implementation manner, the movement speed parameter may include the angular velocity and linear acceleration in the device coordinate system of the target device, that is, the angular velocity and linear acceleration in the body coordinate system (ie, body frame) of the target device. Since the IMU includes three single-axis accelerometers and three single-axis gyroscopes, the accelerometers can be used to detect the respective linear accelerations of the target device on the x-axis, y-axis, and z-axis in the device coordinate system. The gyroscope detects the respective angular velocities of the target device on the x-axis, y-axis, and z-axis in the device coordinate system; it can be understood that the movement speed parameters can include the x-axis, y-axis, and z-axis in the device coordinate system of the target device respective angular velocity and linear acceleration. It should be noted that the IMU may also be a 9-axis IMU, that is, include a magnetometer, and the IMU is not specifically limited in this embodiment.
作为一个示例,为在目标设备丢失视觉信息的情况下依然可以估计目标设备的位姿,本实施例通过IMU获取目标设备在前一时刻的运动速度参数预测目标设备当前时刻的位姿。需要说明的是,该前一时刻为目标设备丢失视觉信息前的一历史时刻。As an example, in order to estimate the pose of the target device even if the target device loses visual information, this embodiment uses the IMU to obtain the motion velocity parameters of the target device at the previous moment to predict the pose of the target device at the current moment. It should be noted that the previous moment is a historical moment before the target device loses the visual information.
需要说明的是,在本实施例的一种实现方式中,在所述获取目标设备在前一时刻的运动速度参数的步骤之前,所述方法还可以包括以下步骤:It should be noted that, in an implementation manner of this embodiment, before the step of acquiring the movement speed parameter of the target device at the previous moment, the method may further include the following steps:
判断是否检测到所述目标设备的视觉信息;judging whether the visual information of the target device is detected;
若未检测到所述目标设备的视觉信息,则执行所述获取目标设备在前一时刻的运动速度参数。If the visual information of the target device is not detected, the acquisition of the movement speed parameter of the target device at the previous moment is performed.
可以理解的是,在本实现方式中,若未检测到目标设备的视觉信息,则说明目标设备的视觉信息丢失,故不可以利用视觉信息估计目标设备的位姿;此时,需获取目标设备在前一时刻的运动速度参数,以便后续利用目标设备的运动速度参数估计目标设备的位姿。It can be understood that in this implementation, if the visual information of the target device is not detected, it means that the visual information of the target device is lost, so it is not possible to use the visual information to estimate the pose of the target device; at this time, it is necessary to obtain the target device The motion speed parameters at the previous moment, so that the motion speed parameters of the target device can be used to estimate the pose of the target device.
更具体地,假设目标设备中搭载了SLAM系统,在SLAM跟踪线程主函数中,当SLAM系统的系统状态为正常时,SLAM系统处于正常跟踪状态;当视觉信息丢失时,SLAM系统进入跟踪丢失状态,此时可利用IMU采集的运动速度参数进行积分运算得到递推的位姿,但是由于通过积分运算所得到的递推位置的可信度不高,因此,在视觉信息丢失超过预设时长(比如1s)之后,视觉信息还未恢复正常,则SLAM系统进入跟踪丢失状态,此时,需执行所述获取目标设备在前一时刻的运动速度参数的步骤。More specifically, assuming that the target device is equipped with a SLAM system, in the main function of the SLAM tracking thread, when the system state of the SLAM system is normal, the SLAM system is in the normal tracking state; when the visual information is lost, the SLAM system enters the tracking loss state , at this time, the motion velocity parameters collected by the IMU can be used to perform integral operations to obtain the recursive pose, but since the reliability of the recursive position obtained through the integral operation is not high, when the loss of visual information exceeds the preset duration ( For example, after 1s), if the visual information has not returned to normal, the SLAM system enters the tracking loss state. At this time, the step of obtaining the motion speed parameter of the target device at the previous moment needs to be executed.
S202:利用运动速度参数和预设位移预测模型,估计目标设备各当前时刻的预测位移增量。S202: Estimate the predicted displacement increment of the target device at each current moment by using the motion speed parameter and the preset displacement prediction model.
其中,当前时刻为目标设备丢失视觉信息时的最新时刻和丢失视觉信息后的每个时刻。Wherein, the current moment is the latest moment when the target device loses the visual information and every moment after the visual information is lost.
在本实施例中,获取到目标设备在历史时刻下的运动速度参数后,可利用预设位移预测模型,根据运动速度参数确定目标设备后续每个当前时刻的预测位移增量,其中,目标设备当前时刻的预测位移增量可以理解为目标设备丢失视觉信息的时间内每个时刻的位移预测变化量。In this embodiment, after obtaining the movement speed parameters of the target device at historical moments, the preset displacement prediction model can be used to determine the subsequent predicted displacement increment of the target device at each current moment according to the movement speed parameters, wherein the target device The predicted displacement increment at the current moment can be understood as the predicted variation of displacement at each moment during the time when the target device loses visual information.
作为一种实施例,可根据目标设备前一时刻的设备坐标系下的角速度,确定目标旋转矩阵。优选地,对目标设备的设备坐标系下的角速度进行积分,得到目标旋转矩阵,其中,目标旋转矩阵可以理解为将运动速度参数由设备坐标系(即body frame)转换到SLAM系统的预设世界坐标系(world frame)下的旋转矩阵(即旋转Rwb)。As an embodiment, the target rotation matrix may be determined according to the angular velocity of the target device in the device coordinate system at the previous moment. Preferably, the angular velocity under the device coordinate system of the target device is integrated to obtain the target rotation matrix, wherein the target rotation matrix can be understood as converting the motion velocity parameter from the device coordinate system (ie body frame) to the preset world of the SLAM system The rotation matrix (that is, the rotation Rwb) in the coordinate system (world frame).
进一步利用目标旋转矩阵对设备坐标系下的角速度和线加速度进行旋转,得到世界坐标系下的角速度和线加速度,并将世界坐标系下的角速度和线加速度输入预设位移预测模型,得到目标设备各当前时刻的预测位移增量。Further use the target rotation matrix to rotate the angular velocity and linear acceleration in the equipment coordinate system to obtain the angular velocity and linear acceleration in the world coordinate system, and input the angular velocity and linear acceleration in the world coordinate system into the preset displacement prediction model to obtain the target equipment Predicted displacement increments at each current moment.
可以理解的是,预设位移预测模型的输入可以为固定时间长度(比如1s)的世界坐标系下的x轴、y轴、z轴的角速度和线加速度;需要说明的是,假设固定时间长度为1s,由于一个时刻的角速度和线加速度的数据为6维,如果角速度和线加速度的采集频率为250HZ,则1s中采集到的角速度和线加速度数据有250个,可以理解的是,预设位移预测模型的输入的张量大小为batch_size×6×250。而预设位移预测模型的输出为目标设备在世界坐标系下的x轴、y轴、z轴的预测位移增量,即在一段时间内,目标设备在世界坐标系下的x轴、y轴、z轴上的位移增量。It can be understood that the input of the preset displacement prediction model can be the angular velocity and linear acceleration of the x-axis, y-axis, and z-axis in the world coordinate system of a fixed time length (such as 1s); it should be noted that, assuming a fixed time length is 1s, since the data of angular velocity and linear acceleration at one moment is 6-dimensional, if the acquisition frequency of angular velocity and linear acceleration is 250HZ, then there are 250 pieces of angular velocity and linear acceleration data collected in 1s. It is understandable that the preset The input tensor size of the displacement prediction model is batch_size×6×250. The output of the preset displacement prediction model is the predicted displacement increment of the x-axis, y-axis, and z-axis of the target device in the world coordinate system, that is, within a period of time, the x-axis and y-axis of the target device in the world coordinate system , The displacement increment on the z axis.
在一个实施例中,预设位移预测模型包括多个级联的卷积层和一个输出层,如图3所示,其中,输出层与多个级联的卷积层中最后一级卷积层相连接,并且,输出层由括全局平均池化层(AngPool1d)和卷积层(Conv1d)串联组成,进而保留网络的空间结构,极大地降低模型的参数量,减小推理时间,同时防 止过拟合的技术效果。需要说明的是,位移预测模型中还包括BN层和relu层,但是在图3中省略不展示,其中,图3中的c代表通道数,k代表卷即可大小。In one embodiment, the preset displacement prediction model includes a plurality of cascaded convolutional layers and an output layer, as shown in FIG. The layers are connected, and the output layer is composed of a global average pooling layer (AngPool1d) and a convolutional layer (Conv1d) in series, thereby retaining the spatial structure of the network, greatly reducing the number of parameters of the model, reducing the reasoning time, and preventing The technical effect of overfitting. It should be noted that the displacement prediction model also includes a BN layer and a relu layer, but they are omitted in Figure 3, where c in Figure 3 represents the number of channels, and k represents the size of the volume.
进一步地,在预先对预设位移预测模型进行模型训练的过程中,可利用损失函数对模型进行监督,其中,损失函数优选为均方误差(Mean-Square Error,MSE)损失函数。MSE LOSS(即MSE损失函数)可以如以下公式所示:Further, in the process of performing model training on the preset displacement prediction model in advance, the model can be supervised by using a loss function, wherein the loss function is preferably a mean-square error (Mean-Square Error, MSE) loss function. MSE LOSS (that is, the MSE loss function) can be shown in the following formula:
Figure PCTCN2022100638-appb-000001
Figure PCTCN2022100638-appb-000001
其中,L mse为MSE LOSS,
Figure PCTCN2022100638-appb-000002
是位移预测模型输出的三维位移增量,
Figure PCTCN2022100638-appb-000003
是三维位移增量的真值(即真实值),i∈[0,n],表示第i个时刻。
Among them, L mse is MSE LOSS,
Figure PCTCN2022100638-appb-000002
is the three-dimensional displacement increment output by the displacement prediction model,
Figure PCTCN2022100638-appb-000003
is the true value (that is, the real value) of the three-dimensional displacement increment, i∈[0,n], representing the i-th moment.
需要说明的是,位移预测模型的网络结构还可以为resnet18的一维形式、TCN、LSTM等其他神经网络结构,此处不作限制。It should be noted that the network structure of the displacement prediction model can also be a one-dimensional form of resnet18, TCN, LSTM and other neural network structures, which are not limited here.
S203:根据目标设备各当前时刻的预测位移增量计算目标设备各当前时刻对应的预测位姿,并构建目标设备的预测运动轨迹。S203: Calculate the predicted pose of the target device corresponding to each current moment according to the predicted displacement increment of the target device at each current moment, and construct a predicted motion trajectory of the target device.
在确定目标设备丢失视觉信息后每个当前时刻的预测位移增量后,可以根据各当前时刻预测位移增量以及目标设备在前一时刻的位姿信息,即得到目标设备在各当前时刻的预测位姿信息,进而利用所有当前时刻的预测位姿信息构建目标设备的预测运动轨迹。After determining the predicted displacement increment at each current moment after the target device loses visual information, the predicted displacement increment at each current moment and the pose information of the target device at the previous moment can be used to obtain the prediction of the target device at each current moment Pose information, and then use all the predicted pose information at the current moment to construct the predicted motion trajectory of the target device.
作为一种示例,目标设备的预测位移增量包括X轴上的预测位移增量、Y轴上的预测位移增量和Z轴上的预测位移增量。优选地,根据目标设备前一时刻在X轴上的预测位移增量、Y轴上的预测位移增量、Z轴上的预测位移增量以及目标设备在前一时刻的位姿信息,确定目标设备当前时刻在X轴上的预测位姿、Y轴上的预测位姿和Z轴上的预测位姿,例如,可以在目标设备在前一时刻的位姿信息(即x轴、y轴、z轴上的坐标)上添加目标设备在X轴上的预测位移增量、Y轴上的预测位移增量、Z轴上的预测位移增量,得到目标设备当前时刻在X轴上的预测位姿、Y轴上的预测位姿和Z轴上的预测位姿(即 预测坐标值)。然后,可以根据目标设备每个前一时刻在X轴上的预测位姿、Y轴上的预测位姿和Z轴上的预测位姿,确定目标设备丢失视觉信息后每个当前时刻的预测位姿信息,即可以将目标设备当前时刻在X轴上的预测位置、Y轴上的预测位置和Z轴上的预测位置作为目标设备当前时刻的预测位姿信息;将目标设备当前时刻的预测位姿信息视为前一时刻的位姿信息,重复上述计算,预测下一当前时刻的预测位姿,从而获取目标设备丢失视觉信息的时间段内每个当前时刻的预测位姿。As an example, the predicted displacement increment of the target device includes a predicted displacement increment on the X axis, a predicted displacement increment on the Y axis, and a predicted displacement increment on the Z axis. Preferably, the target is determined according to the predicted displacement increment on the X-axis, the predicted displacement increment on the Y-axis, the predicted displacement increment on the Z-axis, and the pose information of the target device at the previous moment. The predicted pose of the device on the X-axis, the predicted pose on the Y-axis, and the predicted pose on the Z-axis at the current moment, for example, the pose information of the target device at the previous moment (that is, the x-axis, y-axis, Coordinates on the z-axis) add the predicted displacement increment of the target device on the X-axis, the predicted displacement increment on the Y-axis, and the predicted displacement increment on the Z-axis to obtain the predicted position of the target device on the X-axis at the current moment pose, the predicted pose on the Y axis, and the predicted pose on the Z axis (that is, the predicted coordinate value). Then, according to the predicted pose of the target device on the X-axis, the predicted pose on the Y-axis, and the predicted pose on the Z-axis at each previous moment, the predicted position of the target device at each current moment after losing visual information can be determined Attitude information, that is, the predicted position of the target device on the X-axis, the predicted position on the Y-axis, and the predicted position on the Z-axis at the current moment can be used as the predicted pose information of the target device at the current moment; the predicted position of the target device at the current moment The pose information is regarded as the pose information of the previous moment, and the above calculation is repeated to predict the predicted pose of the next current moment, so as to obtain the predicted pose of each current moment during the time period when the target device loses visual information.
进一步地,在目标设备丢失视觉信息的时间段内,利用每个当前时刻的位移的预测位姿信息构建目标设备的预测运动轨迹,具体如图4的实线所示。Further, during the time period when the target device loses visual information, the predicted motion trajectory of the target device is constructed using the predicted pose information of the displacement at each current moment, as shown by the solid line in FIG. 4 .
应当理解的是,本申请除了利用深度学习方法进行位姿估计外,还可以通过零速更新结合卡尔曼滤波(EKF)、Step Counting等方法进行位姿估计,此处不作限制。It should be understood that, in addition to using deep learning methods for pose estimation, this application can also perform pose estimation through zero-speed update combined with Kalman filter (EKF), Step Counting and other methods, and there is no limitation here.
S204:对目标设备的预测位姿及预测运动轨迹进行优化,获取目标设备各当前时刻的目标位姿及目标运动轨迹。S204: Optimizing the predicted pose and predicted motion trajectory of the target device, and obtaining the target pose and target motion trajectory of the target device at each current moment.
需要说明的是,虽然可以利用目标设备的运动速度参数,估计目标设备的预测位姿,但为了保证所确定的目标设备的运行轨迹的准确性,还需要不断检测是否有采集到目标设备的视觉信息,以便可以根据采集到的目标设备的视觉信息对目标设备的预测运动轨迹进行校正。It should be noted that although the target device's motion velocity parameters can be used to estimate the predicted pose of the target device, in order to ensure the accuracy of the determined running trajectory of the target device, it is necessary to continuously check whether there is a visual image of the target device collected. Information, so that the predicted motion trajectory of the target device can be corrected according to the collected visual information of the target device.
在一种具体实现方式中,步骤S204还包括:In a specific implementation manner, step S204 also includes:
若检测到目标设备的视觉信息且视觉信息满足重定位条件,则根据视觉信息中的图像帧对目标设备的预测位姿进行位姿优化,得到目标位姿。If the visual information of the target device is detected and the visual information satisfies the relocation conditions, the predicted pose of the target device is optimized according to the image frame in the visual information to obtain the target pose.
更具体地,可先判断检测到所采集的目标设备的视觉信息(比如采集到两个图像帧后)是否满足重定位条件,其中,重定位条件可以为可以根据该视觉信息对目标设备进行定位,确定目标设备的目标位姿,即在检测到视觉信息后,判断该视觉信息中的图像帧是否和SLAM系统预先建立的全局地图或者局部地图中的一帧图像帧匹配成功;若该视觉信息中的图像帧和SLAM系统预先建立 的全局地图或者局部地图中的一帧图像帧匹配成功,说明该视觉信息满足重定位条件,从而可以根据该视觉信息对目标设备进行定位,确定目标设备的目标位姿;反之,若该视觉信息中的图像帧和SLAM系统预先建立的全局地图或者局部地图中的任意一帧图像帧均不能匹配成功,则说明该视觉信息不满足重定位条件,从而不能根据该视觉信息对目标设备进行定位确定目标设备的目标位姿。More specifically, it may first be determined whether the detected visual information of the target device (for example, after two image frames are collected) satisfies the relocation condition, wherein the relocation condition may be that the target device can be positioned according to the visual information , determine the target pose of the target device, that is, after detecting the visual information, judge whether the image frame in the visual information matches successfully with an image frame in the global map or local map pre-established by the SLAM system; if the visual information The image frame in the SLAM system is successfully matched with an image frame in the global map or local map pre-established by the SLAM system, indicating that the visual information satisfies the relocation condition, so that the target device can be located according to the visual information and the target of the target device can be determined Conversely, if the image frame in the visual information cannot match any image frame in the pre-established global map or local map of the SLAM system, it means that the visual information does not meet the relocation conditions, so it cannot be based on The visual information locates the target device to determine the target pose of the target device.
在检测到目标设备的视觉信息且确定视觉信息满足重定位条件后,可以根据所述视觉信息中的图像帧对所述目标设备的预测位姿进行位姿优化,以得到目标位姿。After the visual information of the target device is detected and it is determined that the visual information satisfies the relocation condition, a pose optimization may be performed on the predicted pose of the target device according to the image frames in the visual information to obtain a target pose.
作为一种示例,可以先根据所述视觉信息中的图像帧,确定目标设备对应的目标位姿。具体地,可以先在SLAM系统预先建立的全局地图或者局部地图中确定与视觉信息中的图像帧相匹配的图像帧,其中,视觉信息中的图像帧与该相匹配的图像帧的相似度大于预设阈值,或者,该相匹配的图像帧为地图的所有图像帧中与视觉信息中的图像帧的相似度为最大的;然后,可以根据视觉信息中的图像帧以及地图中与其相匹配的图像帧,确定目标设备的理想位姿。As an example, the target pose corresponding to the target device may be determined first according to the image frame in the visual information. Specifically, the image frame that matches the image frame in the visual information can be determined in the global map or local map pre-established in the SLAM system, wherein the similarity between the image frame in the visual information and the matched image frame is greater than Preset threshold, or, the matching image frame is the largest similarity between all image frames in the map and the image frame in the visual information; then, according to the image frame in the visual information and the matching image frame in the map Image frames, determine the ideal pose of the target device.
确定目标设备对应的理想位姿后,根据理想位姿对未检测到视觉信息之前的预测位姿及预测运动轨迹进行矫正优化,获取目标设备的目标位姿与目标运动轨迹。具体地,可先获取未检测到视觉信息之前的预测位姿,即先获取到利用预设位移预测模型确定目标设备的预测位姿;然后根据理想位姿对未检测到视觉信息之前的预测位姿及预测位姿对应的预测运动轨迹进行优化,获取目标设备的目标位姿与目标运动轨迹。优选地,以预测位姿为优化变量,以预测位姿与理想位姿之间相对运动估计(即目标设备的预测位姿对应的预测运动轨迹)的误差为误差项,构建的优化模型,该优化模型如下列公式所示:After the ideal pose corresponding to the target device is determined, the predicted pose and predicted motion trajectory before the visual information is not detected are corrected and optimized according to the ideal pose, and the target pose and target motion trajectory of the target device are obtained. Specifically, the predicted pose before no visual information is detected can be obtained first, that is, the predicted pose of the target device is determined using a preset displacement prediction model; and then the predicted pose before no visual information is detected according to the ideal pose The predicted motion trajectory corresponding to the pose and predicted pose is optimized to obtain the target pose and target motion trajectory of the target device. Preferably, the predicted pose is used as the optimization variable, and the error of the relative motion estimation between the predicted pose and the ideal pose (that is, the predicted motion trajectory corresponding to the predicted pose of the target device) is used as an error term to construct an optimization model, the The optimization model is shown in the following formula:
Figure PCTCN2022100638-appb-000004
Figure PCTCN2022100638-appb-000004
其中,T i是第i时刻目标设备的预测位姿;σ是残差的协方差;
Figure PCTCN2022100638-appb-000005
表 示是残差之和的二范数(最小二乘);min()表示最小化;T={T i} i∈[0,n]表示待优化的预测位姿;ΔT m为预测位姿与理想位姿之间相对位姿的观测值;r()为残差函数,表示待优化的预测位姿与理想位姿之间相对位姿与其观测值之间的残差,具体为:
Figure PCTCN2022100638-appb-000006
接着,对优化模型进行求解,便可以得到目标运动轨迹,其中,目标运动轨迹中视觉信息丢失阶段的起点和终点分别为未丢失视觉信息之前所检测的位姿和目标设备的理想位姿。即经过优化后,目标设备的预测位姿对应的预测运动轨迹的误差会得以减小,并且优化后得到的目标运动轨迹中视觉信息丢失阶段的轨迹的首尾会和SLAM系统根据采集到的视觉信息正常估计的轨迹对齐;例如,如图4所示,视觉信息丢失阶段中的实线线段为预测位姿对应的预测运动轨迹,虚线线段为优化后的目标运动轨迹。根据优化后的目标运动轨迹对所述目标设备的预测位姿进行位姿优化,得到目标位姿,从而SLAM系统可以基于较为准确的绝对位置(即目标位姿)继续跟踪。
Among them, T i is the predicted pose of the target device at the i-th moment; σ is the covariance of the residual;
Figure PCTCN2022100638-appb-000005
Represents the two-norm (least squares) of the sum of residuals; min() represents minimization; T={T i } i∈[0,n] represents the predicted pose to be optimized; ΔT m is the predicted pose The observed value of the relative pose with the ideal pose; r() is a residual function, which represents the residual between the relative pose between the predicted pose to be optimized and the ideal pose and its observed value, specifically:
Figure PCTCN2022100638-appb-000006
Then, the optimization model is solved to obtain the target motion trajectory, where the starting point and end point of the visual information loss stage in the target motion trajectory are the detected pose before the loss of visual information and the ideal pose of the target device, respectively. That is, after optimization, the error of the predicted motion trajectory corresponding to the predicted pose of the target device will be reduced, and the head and tail of the trajectory in the visual information loss stage in the optimized target motion trajectory will be compared with the SLAM system based on the collected visual information. Normal estimated trajectory alignment; for example, as shown in Figure 4, the solid line segment in the visual information loss stage is the predicted motion trajectory corresponding to the predicted pose, and the dashed line segment is the optimized target motion trajectory. Perform pose optimization on the predicted pose of the target device according to the optimized target motion trajectory to obtain the target pose, so that the SLAM system can continue tracking based on a relatively accurate absolute position (ie, the target pose).
需要说明的是,若检测到目标设备的视觉信息,且视觉信息不满足重定位条件,则继续执行所述获取目标设备在前一时刻的运动速度参数,直至检测到目标设备的视觉信息,且视觉信息满足重定位条件。当然,为了避免出现长时间未采集到视觉信息时,SLAM系统无法重启的问题,在本实施例的一种实现方式中,若在预设时长内(比如20s)未检测到所述目标设备的视觉信息,则选择系统重启。It should be noted that if the visual information of the target device is detected, and the visual information does not satisfy the relocation condition, then continue to perform the acquisition of the movement speed parameter of the target device at the previous moment until the visual information of the target device is detected, and The visual information satisfies the relocation criteria. Of course, in order to avoid the problem that the SLAM system cannot be restarted when the visual information is not collected for a long time, in an implementation of this embodiment, if the target device is not detected within a preset time period (such as 20s), visual information, select System Reboot.
上述所有可选技术方案,可以采用任意结合形成本申请的可选实施例,在此不再一一赘述。All the above optional technical solutions may be combined in any way to form optional embodiments of the present application, which will not be repeated here.
下述为本申请装置实施例,可以用于执行本申请方法实施例。对于本申请装置实施例中未披露的细节,请参照本申请方法实施例。The following are device embodiments of the present application, which can be used to implement the method embodiments of the present application. For details not disclosed in the device embodiments of the present application, please refer to the method embodiments of the present application.
图5是本申请实施例提供的位姿预测装置的示意图。如图5所示,该位姿预测装置包括:Fig. 5 is a schematic diagram of a pose prediction device provided by an embodiment of the present application. As shown in Figure 5, the pose prediction device includes:
参数获取模块501,用于获取目标设备前一时刻的运动速度参数;其中, 运动速度参数包括目标设备的角速度和线加速度,前一时刻为所述目标设备丢失视觉信息前的一历史时刻;The parameter acquisition module 501 is configured to acquire the movement speed parameter of the target device at a previous moment; wherein, the movement speed parameter includes the angular velocity and linear acceleration of the target device, and the previous moment is a historical moment before the target device loses visual information;
增量计算模块502,用于利用运动速度参数和预设位移预测模型,估计目标设备各当前时刻的预测位移增量;其中,当前时刻为目标设备丢失视觉信息时的最新时刻及丢失视觉信息后的每个时刻;The incremental calculation module 502 is used to estimate the predicted displacement increment of the target device at each current moment by using the motion speed parameter and the preset displacement prediction model; wherein, the current moment is the latest moment when the target device loses visual information and after the visual information is lost every moment of
位姿与轨迹预测模块503,用于根据目标设备各当前时刻的预测位移增量计算目标设备各当前时刻对应的预测位姿,并构建目标设备的预测运动轨迹;The pose and trajectory prediction module 503 is used to calculate the predicted pose of the target device corresponding to each current moment according to the predicted displacement increment of each current moment of the target device, and construct the predicted motion trajectory of the target device;
优化模块504,用于对目标设备的所述预测位姿及预测运动轨迹进行优化,获取目标设备各当前时刻的目标位姿及对应的目标运动轨迹。The optimization module 504 is configured to optimize the predicted pose and predicted motion trajectory of the target device, and obtain the target pose and corresponding target motion trajectory of the target device at each current moment.
可选的,增量计算模块502,具体用于:Optionally, the incremental calculation module 502 is specifically used for:
根据目标设备前一时刻的设备坐标系下的角速度,确定目标旋转矩阵;Determine the target rotation matrix according to the angular velocity of the target device in the device coordinate system at the previous moment;
利用目标旋转矩阵,将设备坐标系下的角速度和线加速度转换为世界坐标系下的角速度和线加速度;Use the target rotation matrix to convert the angular velocity and linear acceleration in the device coordinate system to the angular velocity and linear acceleration in the world coordinate system;
将世界坐标系下的角速度和线加速度输入预设位移预测模型,得到目标设备各当前时刻的预测位移增量。Input the angular velocity and linear acceleration in the world coordinate system into the preset displacement prediction model to obtain the predicted displacement increment of the target equipment at each current moment.
可选的,目标设备的预测位移增量包括X轴上的预测位移增量、Y轴上的预测位移增量和Z轴上的预测位移增量;Optionally, the predicted displacement increment of the target device includes a predicted displacement increment on the X axis, a predicted displacement increment on the Y axis, and a predicted displacement increment on the Z axis;
位置确定模块503,具体用于:The position determination module 503 is specifically used for:
根据所述目标设备各当前时刻在X轴上的预测位移增量、Y轴上的预测位移增量、Z轴上的预测位移增量以及目标设备在前一时刻的位置信息,确定目标设备的在X轴上的预测位置、Y轴上的预测位置和Z轴上的预测位置;According to the predicted displacement increment on the X-axis, the predicted displacement increment on the Y-axis, the predicted displacement increment on the Z-axis of the target device at each current moment, and the position information of the target device at the previous moment, determine the target device's position The predicted position on the X axis, the predicted position on the Y axis, and the predicted position on the Z axis;
根据目标设备的在X轴上的预测位置、Y轴上的预测位置和Z轴上的预测位置,确定目标设备各当前时刻的预测位姿。According to the predicted position of the target device on the X-axis, the predicted position on the Y-axis and the predicted position on the Z-axis, the predicted poses of the target device at each current moment are determined.
可选的,装置还包括检测模块;检测模块,用于:Optionally, the device also includes a detection module; the detection module is used for:
判断是否检测到目标设备的视觉信息;Judging whether the visual information of the target device is detected;
若未检测到目标设备的视觉信息,则执行所述获取目标设备在前一时刻的 运动速度参数并执行上述位姿预测方法。If the visual information of the target device is not detected, then perform the acquisition of the motion speed parameter of the target device at the previous moment and perform the above-mentioned pose prediction method.
可选的,优化模块,具体用于:Optional, optimization module, specifically for:
若检测到目标设备的视觉信息满足重定位条件,则根据视觉信息中的图像帧对目标设备的预测位姿及预测运动轨迹进行优化,得到目标位姿及目标运动轨迹。If it is detected that the visual information of the target device satisfies the relocation conditions, the predicted pose and predicted motion trajectory of the target device are optimized according to the image frames in the visual information to obtain the target pose and target motion trajectory.
可选的,优化模块,具体用于:Optional, optimization module, specifically for:
根据视觉信息中的图像帧,确定目标设备对应的理想位姿;According to the image frame in the visual information, determine the ideal pose corresponding to the target device;
根据理想位姿与未检测到视觉信息之前所检测的预测位姿对预测运动轨迹进行优化,确定预测运动轨迹;Optimize the predicted motion trajectory according to the ideal pose and the predicted pose detected before the visual information is not detected, and determine the predicted motion trajectory;
根据目标运动轨迹对目标设备的预测位姿进行位姿优化,得到目标位姿。According to the target motion trajectory, the predicted pose of the target device is optimized to obtain the target pose.
可选的,优化模块,还用于:Optionally, the optimization module is also used to:
若检测到目标设备的视觉信息不满足重定位条件,则继续执行获取目标设备在前一时刻的运动速度参数。If it is detected that the visual information of the target device does not satisfy the relocation condition, continue to obtain the movement speed parameter of the target device at the previous moment.
可选的,优化模块,还用于:Optionally, the optimization module is also used to:
若在预设时长内未检测到目标设备的视觉信息,则重启。If the visual information of the target device is not detected within the preset time period, it will restart.
应理解,上述实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实施过程构成任何限定。It should be understood that the sequence numbers of the steps in the above embodiments do not mean the order of execution, and the execution order of each process should be determined by its function and internal logic, and should not constitute any limitation to the implementation process of the embodiment of the present application.
图6是本申请实施例提供的终端设备6的示意图。终端设备6包括:处理器601、存储器602以及存储在该存储器602中并且可以在处理器601上运行的计算机程序603。处理器601执行计算机程序603时实现上述各个方法实施例中的步骤。或者,处理器601执行计算机程序603时实现上述各装置实施例中各模块/模块的功能。FIG. 6 is a schematic diagram of a terminal device 6 provided by an embodiment of the present application. The terminal device 6 includes: a processor 601 , a memory 602 and a computer program 603 stored in the memory 602 and capable of running on the processor 601 . When the processor 601 executes the computer program 603, the steps in the foregoing method embodiments are implemented. Alternatively, when the processor 601 executes the computer program 603, the functions of the modules/modules in the foregoing device embodiments are realized.
示例性地,计算机程序603可以被分割成一个或多个模块/模块,一个或多个模块/模块被存储在存储器602中,并由处理器601执行,以完成本申请。一个或多个模块/模块可以是能够完成特定功能的一系列计算机程序指令段,该指 令段用于描述计算机程序603在计算机设备6中的执行过程。Exemplarily, the computer program 603 can be divided into one or more modules/modules, and one or more modules/modules are stored in the memory 602 and executed by the processor 601 to complete the present application. One or more modules/modules may be a series of computer program instruction segments capable of performing specific functions, and the instruction segments are used to describe the execution process of the computer program 603 in the computer device 6.
终端设备6可以包括但不仅限于处理器601和存储器602。本领域技术人员可以理解,图6仅仅是终端设备6的示例,并不构成对终端设备6的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如,计算机设备还可以包括输入输出设备、网络接入设备、总线等。The terminal device 6 may include, but not limited to, a processor 601 and a memory 602 . Those skilled in the art can understand that FIG. 6 is only an example of the terminal device 6, and does not constitute a limitation on the terminal device 6. It may include more or less components than those shown in the figure, or combine certain components, or different components. , for example, computer equipment may also include input and output equipment, network access equipment, bus, and so on.
处理器601可以是中央处理模块(Central Processing Unit,CPU),也可以是其它通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现场可编程门阵列(Field-Programmable Gate Array,FPGA)或者其它可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The processor 601 can be a central processing unit (Central Processing Unit, CPU), or other general processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), on-site Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc. A general-purpose processor may be a microprocessor, or the processor may be any conventional processor, or the like.
存储器602可以是终端设备6的内部存储模块,例如,终端设备6的硬盘或内存。存储器602也可以是计算机设备6的外部存储设备,例如,终端设备6上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,存储器602还可以既包括终端设备6的内部存储模块也包括外部存储设备。存储器602用于存储计算机程序以及计算机设备所需的其它程序和数据。存储器602还可以用于暂时地存储已经输出或者将要输出的数据。The storage 602 may be an internal storage module of the terminal device 6 , for example, a hard disk or memory of the terminal device 6 . The memory 602 can also be an external storage device of the computer device 6, for example, a plug-in hard disk equipped on the terminal device 6, a smart memory card (Smart Media Card, SMC), a secure digital (Secure Digital, SD) card, a flash memory card ( Flash Card), etc. Further, the memory 602 may also include both an internal storage module of the terminal device 6 and an external storage device. The memory 602 is used to store computer programs and other programs and data required by the computer equipment. The memory 602 can also be used to temporarily store data that has been output or will be output.
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能模块、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能模块、模块完成,即将装置的内部结构划分成不同的功能模块或模块,以完成以上描述的全部或者部分功能。实施例中的各功能模块、模块可以集成在一个处理模块中,也可以是各个模块单独物理存在,也可以两个或两个以上模块集成在一个模块中,上述集成的模块既可以采用硬件的形式实现,也可以采用软件功能模块的形式实现。另外,各功能模块、模块的具体名称也只是为了便于相互区分,并不用于限制本申请的保护范围。上述系 统中模块、模块的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。Those skilled in the art can clearly understand that for the convenience and brevity of description, only the above functional modules and the division of modules are used for illustration. In practical applications, the above functions can be assigned to different functional modules, Module completion means that the internal structure of the device is divided into different functional modules or modules to complete all or part of the functions described above. The functional modules and modules in the embodiments can be integrated into one processing module, or each module can exist separately physically, or two or more modules can be integrated into one module, and the above-mentioned integrated modules can either use hardware It can also be implemented in the form of software function modules. In addition, the specific names of the functional modules and modules are only for the convenience of distinguishing each other, and are not used to limit the protection scope of the present application. For the specific working process of modules and modules in the above-mentioned system, reference may be made to the corresponding process in the aforementioned method embodiments, which will not be repeated here.
在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述或记载的部分,可以参见其它实施例的相关描述。In the above-mentioned embodiments, the descriptions of each embodiment have their own emphases, and for parts that are not detailed or recorded in a certain embodiment, refer to the relevant descriptions of other embodiments.
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的模块及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。Those skilled in the art can appreciate that the modules and algorithm steps of the examples described in conjunction with the embodiments disclosed herein can be implemented by electronic hardware, or a combination of computer software and electronic hardware. Whether these functions are executed by hardware or software depends on the specific application and design constraints of the technical solution. Skilled artisans may use different methods to implement the described functions for each specific application, but such implementation should not be regarded as exceeding the scope of the present application.
在本申请所提供的实施例中,应该理解到,所揭露的装置/计算机设备和方法,可以通过其它的方式实现。例如,以上所描述的装置/计算机设备实施例仅仅是示意性的,例如,模块或模块的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,多个模块或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通讯连接可以是通过一些接口,装置或模块的间接耦合或通讯连接,可以是电性,机械或其它的形式。In the embodiments provided in this application, it should be understood that the disclosed apparatus/computer equipment and methods can be implemented in other ways. For example, the device/computer device embodiments described above are only illustrative, for example, the division of modules or modules is only a logical function division, and there may be other division methods in actual implementation, and multiple modules or components can be Incorporation may either be integrated into another system, or some features may be omitted, or not implemented. In another point, the mutual coupling or direct coupling or communication connection shown or discussed may be through some interfaces, and the indirect coupling or communication connection of devices or modules may be in electrical, mechanical or other forms.
作为分离部件说明的模块可以是或者也可以不是物理上分开的,作为模块显示的部件可以是或者也可以不是物理模块,即可以位于一个地方,或者也可以分布到多个网络模块上。可以根据实际的需要选择其中的部分或者全部模块来实现本实施例方案的目的。A module described as a separate component may or may not be physically separated, and a component shown as a module may or may not be a physical module, that is, it may be located in one place, or may also be distributed to multiple network modules. Part or all of the modules can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
另外,在本申请各个实施例中的各功能模块可以集成在一个处理模块中,也可以是各个模块单独物理存在,也可以两个或两个以上模块集成在一个模块中。上述集成的模块既可以采用硬件的形式实现,也可以采用软件功能模块的形式实现。In addition, each functional module in each embodiment of the present application may be integrated into one processing module, each module may exist separately physically, or two or more modules may be integrated into one module. The above-mentioned integrated modules can be implemented in the form of hardware or in the form of software function modules.
集成的模块/模块如果以软件功能模块的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读存储介质中。基于这样的理解,本申请 实现上述实施例方法中的全部或部分流程,也可以通过计算机程序来指令相关的硬件来完成,计算机程序可以存储在计算机可读存储介质中,该计算机程序在被处理器执行时,可以实现上述各个方法实施例的步骤。计算机程序可以包括计算机程序代码,计算机程序代码可以为源代码形式、对象代码形式、可执行文件或某些中间形式等。计算机可读介质可以包括:能够携带计算机程序代码的任何实体或装置、记录介质、U盘、移动硬盘、磁碟、光盘、计算机存储器、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、电载波信号、电信信号以及软件分发介质等。需要说明的是,计算机可读介质包含的内容可以根据司法管辖区内立法和专利实践的要求进行适当的增减,例如,在某些司法管辖区,根据立法和专利实践,计算机可读介质不包括电载波信号和电信信号。If the integrated modules/modules are implemented in the form of software function modules and sold or used as independent products, they can be stored in a computer-readable storage medium. Based on this understanding, the present application realizes all or part of the processes in the methods of the above embodiments, and can also be completed by instructing related hardware through computer programs. The computer programs can be stored in computer-readable storage media, and the computer programs can be processed. When executed by the controller, the steps in the above-mentioned method embodiments can be realized. A computer program may include computer program code, which may be in source code form, object code form, executable file, or some intermediate form or the like. The computer-readable medium may include: any entity or device capable of carrying computer program code, recording medium, U disk, removable hard disk, magnetic disk, optical disk, computer memory, read-only memory (Read-Only Memory, ROM), random access Memory (Random Access Memory, RAM), electrical carrier signal, telecommunication signal and software distribution medium, etc. It should be noted that the content contained in computer readable media may be appropriately increased or decreased according to the requirements of legislation and patent practice in the jurisdiction. For example, in some jurisdictions, computer readable media may not Including electrical carrier signals and telecommunication signals.
以上实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。The above embodiments are only used to illustrate the technical solutions of the present application, rather than to limit them; although the present application has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that: it can still apply to the foregoing embodiments Modifications to the technical solutions recorded, or equivalent replacements for some of the technical features; and these modifications or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions of each embodiment of the application, and should be included in this application. within the scope of protection.

Claims (12)

  1. 一种位姿预测方法,其特征在于,所述方法包括:A pose prediction method, is characterized in that, described method comprises:
    获取目标设备前一时刻的运动速度参数;其中,所述运动速度参数包括目标设备的角速度和线加速度,所述前一时刻为所述目标设备丢失视觉信息前的一历史时刻;Obtaining the movement speed parameter of the target device at a previous moment; wherein the movement speed parameter includes the angular velocity and linear acceleration of the target device, and the previous moment is a historical moment before the target device loses visual information;
    利用所述运动速度参数和预设位移预测模型,估计所述目标设备各当前时刻的预测位移增量;其中,所述当前时刻为所述目标设备丢失视觉信息时的最新时刻及丢失视觉信息后的每个时刻;Estimate the predicted displacement increment of the target device at each current moment by using the motion speed parameter and the preset displacement prediction model; wherein, the current moment is the latest moment when the target device loses visual information and after losing visual information every moment of
    根据所述目标设备各当前时刻的所述预测位移增量计算所述目标设备各当前时刻对应的预测位姿,并构建所述目标设备的预测运动轨迹;calculating the predicted pose corresponding to each current moment of the target device according to the predicted displacement increment of the target device at each current moment, and constructing a predicted motion trajectory of the target device;
    对所述目标设备的所述预测位姿及所述预测运动轨迹进行优化,获取所述目标设备各当前时刻的目标位姿及对应的目标运动轨迹。Optimizing the predicted pose and the predicted motion trajectory of the target device, and acquiring the target pose and corresponding target motion trajectory of the target device at each current moment.
  2. 根据权利要求1所述的方法,其特征在于,所述利用所述运动速度参数和预设位移预测模型,估计所述目标设备各当前时刻的预测位移增量,包括:The method according to claim 1, wherein the estimating the predicted displacement increment of the target device at each current moment by using the motion speed parameter and the preset displacement prediction model includes:
    根据所述目标设备前一时刻的设备坐标系下的角速度,确定目标旋转矩阵;Determine the target rotation matrix according to the angular velocity of the target device in the device coordinate system at the previous moment;
    利用所述目标旋转矩阵,将所述设备坐标系下的角速度和线加速度转换为世界坐标系下的角速度和线加速度;Using the target rotation matrix, converting the angular velocity and linear acceleration in the device coordinate system into angular velocity and linear acceleration in the world coordinate system;
    将所述世界坐标系下的角速度和线加速度输入所述预设位移预测模型,得到所述目标设备各当前时刻的预测位移增量。Inputting the angular velocity and linear acceleration in the world coordinate system into the preset displacement prediction model to obtain the predicted displacement increment of the target device at each current moment.
  3. 根据权利要求2所述的方法,其特征在于,所述目标设备的预测位移增量包括X轴上的预测位移增量、Y轴上的预测位移增量和Z轴上的预测位移增量;The method according to claim 2, wherein the predicted displacement increment of the target device includes a predicted displacement increment on the X axis, a predicted displacement increment on the Y axis, and a predicted displacement increment on the Z axis;
    所述根据所述目标设备各当前时刻的所述预测位移增量计算所述目标设备各当前时刻对应的预测位姿,包括:The calculating the predicted pose corresponding to each current moment of the target device according to the predicted displacement increment of the target device at each current moment includes:
    根据所述目标设备各当前时刻在X轴上的预测位移增量、Y轴上的预测位 移增量、Z轴上的预测位移增量以及所述目标设备在前一时刻的位置信息,确定所述目标设备的在X轴上的预测位置、Y轴上的预测位置和Z轴上的预测位置;According to the predicted displacement increment on the X-axis, the predicted displacement increment on the Y-axis, the predicted displacement increment on the Z-axis of the target device at each current moment, and the position information of the target device at the previous moment, determine the The predicted position on the X-axis, the predicted position on the Y-axis and the predicted position on the Z-axis of the target device;
    根据所述目标设备的在X轴上的预测位置、Y轴上的预测位置和Z轴上的预测位置,确定所述目标设备各当前时刻的预测位姿。According to the predicted position of the target device on the X-axis, the predicted position on the Y-axis, and the predicted position on the Z-axis, the predicted poses of the target device at each current moment are determined.
  4. 根据权利要求1至3中任一所述的方法,其特征在于,所述预设位移预测模型包括多个级联的卷积层和一个输出层;其中,所述输出层与所述多个级联的卷积层中最后一级卷积层相连接,并且,所述输出层由全局平均池化层和卷积层串联组成;所述位移预测模型的损失函数为均方误差损失函数。The method according to any one of claims 1 to 3, wherein the preset displacement prediction model includes a plurality of cascaded convolutional layers and an output layer; wherein, the output layer and the plurality of The last level of convolutional layers in the cascaded convolutional layers is connected, and the output layer is composed of a global average pooling layer and a convolutional layer in series; the loss function of the displacement prediction model is a mean square error loss function.
  5. 根据权利要求1至3中任一所述的方法,其特征在于,在所述获取目标设备在前一时刻的运动速度参数的步骤之前,所述方法还包括:The method according to any one of claims 1 to 3, characterized in that, before the step of acquiring the movement speed parameter of the target device at the previous moment, the method further comprises:
    判断是否检测到所述目标设备的视觉信息;judging whether the visual information of the target device is detected;
    若未检测到所述目标设备的视觉信息,则执行所述获取目标设备在前一时刻的运动速度参数的步骤。If the visual information of the target device is not detected, the step of acquiring the movement speed parameter of the target device at the previous moment is executed.
  6. 根据权利要求5所述的方法,其特征在于,所述对所述目标设备的所述预测位姿及所述预测运动轨迹进行优化,获取所述目标设备各当前时刻的目标位姿及对应的目标运动轨迹,还包括:The method according to claim 5, characterized in that, optimizing the predicted pose and the predicted motion trajectory of the target device, and obtaining the target pose and the corresponding target pose of the target device at each current moment The target movement trajectory also includes:
    判断所述目标设备的视觉信息是否满足重定位条件,若满足,则根据所述视觉信息中的图像帧对所述目标设备的预测位姿及预测运动轨迹进行优化,得到目标位姿及目标运动轨迹;若不满足,则继续执行所述获取目标设备在前一时刻的运动速度参数的步骤;其中,所述重定位条件为检测到所述视觉信息后,判断所述视觉信息中的图像帧是否和预先建立的全局地图或者局部地图中的一帧图像帧匹配成功。Judging whether the visual information of the target device satisfies the relocation condition, and if so, optimizing the predicted pose and predicted motion trajectory of the target device according to the image frames in the visual information to obtain the target pose and target motion trajectory; if not satisfied, continue to execute the step of acquiring the motion speed parameter of the target device at the previous moment; wherein, the relocation condition is to judge the image frame in the visual information after the visual information is detected Whether it is successfully matched with an image frame in the pre-established global map or local map.
  7. 根据权利要求6所述的方法,其特征在于,所述对所述目标设备的所述预测位姿及所述预测运动轨迹进行优化,获取所述目标设备各当前时刻的目标位姿及对应的目标运动轨迹,包括:The method according to claim 6, characterized in that, optimizing the predicted pose and the predicted motion trajectory of the target device, and obtaining the target pose and the corresponding target pose of the target device at each current moment Target trajectory, including:
    根据所述视觉信息中的图像帧,确定目标设备各当前时刻对应的理想位姿;According to the image frame in the visual information, determine the ideal pose corresponding to each current moment of the target device;
    根据所述理想位姿与未检测到视觉信息之前所估计的预测位姿对所述预测运动轨迹进行优化,确定所述目标运动轨迹;Optimizing the predicted motion trajectory according to the ideal pose and the predicted pose estimated before the visual information is not detected, and determining the target motion trajectory;
    根据所述目标运动轨迹对所述目标设备各当前时刻的预测位姿进行位姿优化,得到所述目标设备各当前时刻的目标位姿。Perform pose optimization on the predicted poses of the target device at each current moment according to the target motion trajectory to obtain the target poses of the target device at each current moment.
  8. 根据权利要求1至3中任一所述的方法,其特征在于,在所述根据所述目标设备各当前时刻的所述预测位移增量计算所述目标设备各当前时刻对应的预测位姿,并构建所述目标设备的预测运动轨迹的步骤之后,所述方法还包括:The method according to any one of claims 1 to 3, characterized in that, calculating the predicted pose corresponding to each current moment of the target device according to the predicted displacement increment of the target device at each current moment, And after the step of constructing the predicted motion trajectory of the target device, the method further includes:
    若在预设时长内未检测到所述目标设备的视觉信息,则重启系统。If the visual information of the target device is not detected within the preset time period, the system is restarted.
  9. 一种位姿预测装置,其特征在于,所述装置包括:A pose prediction device, characterized in that said device comprises:
    参数获取模块,获取目标设备前一时刻的运动速度参数;其中,所述运动速度参数包括目标设备的角速度和线加速度,所述前一时刻为所述目标设备丢失视觉信息前的一历史时刻;A parameter acquisition module, which acquires a movement speed parameter of the target device at a previous moment; wherein, the movement speed parameter includes the angular velocity and linear acceleration of the target device, and the previous moment is a historical moment before the target device loses visual information;
    增量确定模块,用于利用所述运动速度参数和预设位移预测模型,估计所述目标设备各当前时刻的预测位移增量;其中,所述当前时刻为所述目标设备丢失视觉信息时的最新时刻及丢失视觉信息后的每个时刻;An increment determination module, configured to estimate the predicted displacement increment of the target device at each current moment by using the motion speed parameter and the preset displacement prediction model; wherein, the current moment is when the target device loses visual information The latest moment and every moment after the loss of visual information;
    位姿与轨迹预测模块,用于根据所述目标设备各当前时刻的所述预测位移增量计算所述目标设备各当前时刻对应的预测位姿,并构建所述目标设备的预测运动轨迹;A pose and trajectory prediction module, configured to calculate a predicted pose of the target device corresponding to each current moment according to the predicted displacement increment of the target device at each current moment, and construct a predicted motion trajectory of the target device;
    优化模块,用于对所述目标设备的所述预测位姿及所述预测运动轨迹进行优化,获取所述目标设备各当前时刻的目标位姿及对应的目标运动轨迹。An optimization module, configured to optimize the predicted pose and the predicted motion trajectory of the target device, and acquire target poses and corresponding target motion trajectories of the target device at each current moment.
  10. 一种位姿预测系统,其特征在于,所述系统包括惯导传感器及终端设备,其中,所述惯导传感器用于所述目标设备在前一时刻的运动速度参数,所述终端设备用于执行如权利要求1至8中任一项所述方法的步骤。A pose prediction system, characterized in that the system includes an inertial navigation sensor and a terminal device, wherein the inertial navigation sensor is used for the movement speed parameter of the target device at the previous moment, and the terminal device is used for Carrying out the steps of the method as claimed in any one of claims 1 to 8.
  11. 一种终端设备,包括存储器、处理器以及存储在所述存储器中并且可以在所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算 机程序时实现如权利要求1至8中任一项所述方法的步骤。A terminal device, comprising a memory, a processor, and a computer program stored in the memory and capable of running on the processor, characterized in that, when the processor executes the computer program, the following claims 1 to 1 are implemented. 8. The steps of any one of the methods.
  12. 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1至8中任一项所述方法的步骤。A computer-readable storage medium storing a computer program, wherein the computer program implements the steps of the method according to any one of claims 1 to 8 when executed by a processor.
PCT/CN2022/100638 2022-01-26 2022-06-23 Pose prediction method and apparatus WO2023142353A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210093752.1 2022-01-26
CN202210093752.1A CN114593735A (en) 2022-01-26 2022-01-26 Pose prediction method and device

Publications (1)

Publication Number Publication Date
WO2023142353A1 true WO2023142353A1 (en) 2023-08-03

Family

ID=81805920

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/100638 WO2023142353A1 (en) 2022-01-26 2022-06-23 Pose prediction method and apparatus

Country Status (2)

Country Link
CN (1) CN114593735A (en)
WO (1) WO2023142353A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114593735A (en) * 2022-01-26 2022-06-07 奥比中光科技集团股份有限公司 Pose prediction method and device

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110850455A (en) * 2019-10-18 2020-02-28 浙江天尚元科技有限公司 Track recording method based on differential GPS and vehicle kinematics model
CN111161337A (en) * 2019-12-18 2020-05-15 南京理工大学 Accompanying robot synchronous positioning and composition method in dynamic environment
US20210142488A1 (en) * 2019-11-12 2021-05-13 Naver Labs Corporation Method and system for tracking trajectory based on visual localization and odometry
CN113031436A (en) * 2021-02-25 2021-06-25 西安建筑科技大学 Mobile robot model prediction trajectory tracking control system and method based on event triggering
CN113483755A (en) * 2021-07-09 2021-10-08 北京易航远智科技有限公司 Multi-sensor combined positioning method and system based on non-global consistent map
CN114593735A (en) * 2022-01-26 2022-06-07 奥比中光科技集团股份有限公司 Pose prediction method and device

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111489393B (en) * 2019-01-28 2023-06-02 速感科技(北京)有限公司 VSLAM method, controller and mobile device
CN110009681B (en) * 2019-03-25 2021-07-30 中国计量大学 IMU (inertial measurement unit) assistance-based monocular vision odometer pose processing method
CN110195592B (en) * 2019-04-30 2021-02-05 华中科技大学 Intelligent shield tunneling pose prediction method and system based on hybrid deep learning
CN111508024A (en) * 2019-06-27 2020-08-07 浙江大学 Method for estimating pose of robot based on deep learning
CN110702107A (en) * 2019-10-22 2020-01-17 北京维盛泰科科技有限公司 Monocular vision inertial combination positioning navigation method
KR102508318B1 (en) * 2020-02-06 2023-03-10 연세대학교 산학협력단 Estimating method for physical quantities calculatable by integration using artificial neural network
US11435185B2 (en) * 2020-02-21 2022-09-06 Microsoft Technology Licensing, Llc Systems and methods for deep learning-based pedestrian dead reckoning for exteroceptive sensor-enabled devices
CN113945206B (en) * 2020-07-16 2024-04-19 北京图森未来科技有限公司 Positioning method and device based on multi-sensor fusion
CN112556719B (en) * 2020-11-27 2022-01-21 广东电网有限责任公司肇庆供电局 Visual inertial odometer implementation method based on CNN-EKF
CN113065289B (en) * 2021-04-27 2023-04-07 深圳市商汤科技有限公司 Pose prediction method and device, electronic equipment and storage medium
CN113252031A (en) * 2021-04-28 2021-08-13 燕山大学 NARX neural network assisted integrated navigation method
CN113252060B (en) * 2021-05-31 2021-09-21 智道网联科技(北京)有限公司 Vehicle track calculation method and device based on neural network model

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110850455A (en) * 2019-10-18 2020-02-28 浙江天尚元科技有限公司 Track recording method based on differential GPS and vehicle kinematics model
US20210142488A1 (en) * 2019-11-12 2021-05-13 Naver Labs Corporation Method and system for tracking trajectory based on visual localization and odometry
CN111161337A (en) * 2019-12-18 2020-05-15 南京理工大学 Accompanying robot synchronous positioning and composition method in dynamic environment
CN113031436A (en) * 2021-02-25 2021-06-25 西安建筑科技大学 Mobile robot model prediction trajectory tracking control system and method based on event triggering
CN113483755A (en) * 2021-07-09 2021-10-08 北京易航远智科技有限公司 Multi-sensor combined positioning method and system based on non-global consistent map
CN114593735A (en) * 2022-01-26 2022-06-07 奥比中光科技集团股份有限公司 Pose prediction method and device

Also Published As

Publication number Publication date
CN114593735A (en) 2022-06-07

Similar Documents

Publication Publication Date Title
US10852139B2 (en) Positioning method, positioning device, and robot
CN107687850B (en) Unmanned aerial vehicle pose estimation method based on vision and inertia measurement unit
CN112179330B (en) Pose determination method and device of mobile equipment
CN110490900B (en) Binocular vision positioning method and system under dynamic environment
CN112734852B (en) Robot mapping method and device and computing equipment
US20210183100A1 (en) Data processing method and apparatus
CN111209978B (en) Three-dimensional visual repositioning method and device, computing equipment and storage medium
CN108564657B (en) Cloud-based map construction method, electronic device and readable storage medium
WO2022193508A1 (en) Method and apparatus for posture optimization, electronic device, computer-readable storage medium, computer program, and program product
WO2020221307A1 (en) Method and device for tracking moving object
CN111932616B (en) Binocular vision inertial odometer method accelerated by utilizing parallel computation
CN111915675B (en) Particle drift-based particle filtering point cloud positioning method, device and system thereof
CN110942474B (en) Robot target tracking method, device and storage medium
CN111623773B (en) Target positioning method and device based on fisheye vision and inertial measurement
CN114136315B (en) Monocular vision-based auxiliary inertial integrated navigation method and system
CN114323033B (en) Positioning method and equipment based on lane lines and feature points and automatic driving vehicle
WO2023142353A1 (en) Pose prediction method and apparatus
WO2020014864A1 (en) Pose determination method and device, and computer readable storage medium
CN112967340A (en) Simultaneous positioning and map construction method and device, electronic equipment and storage medium
CN115164936A (en) Global pose correction method and device for point cloud splicing in high-precision map manufacturing
CN112556699B (en) Navigation positioning method and device, electronic equipment and readable storage medium
CN113570716A (en) Cloud three-dimensional map construction method, system and equipment
CN116958452A (en) Three-dimensional reconstruction method and system
CN115727871A (en) Track quality detection method and device, electronic equipment and storage medium
CN113034538B (en) Pose tracking method and device of visual inertial navigation equipment and visual inertial navigation equipment

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22923162

Country of ref document: EP

Kind code of ref document: A1