WO2020238790A1 - Camera positioning - Google Patents
Camera positioning Download PDFInfo
- Publication number
- WO2020238790A1 WO2020238790A1 PCT/CN2020/091768 CN2020091768W WO2020238790A1 WO 2020238790 A1 WO2020238790 A1 WO 2020238790A1 CN 2020091768 W CN2020091768 W CN 2020091768W WO 2020238790 A1 WO2020238790 A1 WO 2020238790A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- camera
- processed
- pose
- pixel
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
- G06T7/74—Determining position or orientation of objects or cameras using feature-based methods involving reference images or patches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/80—Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20076—Probabilistic image processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
Definitions
- the present disclosure relates to the field of computer vision, in particular to a camera positioning method and device, and storage medium.
- Visual positioning has a wide range of applications. In the actual application environment, factors such as object movement may affect the accuracy of visual positioning, and even directly cause visual positioning to fail.
- the present disclosure provides a camera positioning method, device, and storage medium.
- a camera positioning method including:
- the absolute pose of the camera that collects the image to be processed in the world coordinate system is determined according to the target image.
- a camera positioning device including:
- An acquisition module for acquiring the prior probability of a movable object appearing at each of the multiple pixels included in the image template
- the execution module is configured to perform an operation of discarding some pixels for an image to be processed that is as large as the image template according to the prior probability to obtain a target image;
- the positioning module is configured to determine, according to the target image, the absolute pose of the camera that collects the image to be processed in the world coordinate system.
- a computer-readable storage medium stores a computer program, and the computer program is used to execute the camera positioning method described in the first aspect.
- a camera positioning device comprising: a processor; and a memory for storing executable instructions of the processor.
- the processor is configured to call executable instructions stored in the memory to implement the camera positioning method described in the first aspect.
- the prior probability that a movable object appears at each of the multiple pixels included in the image template can be obtained first, and some pixels are discarded based on the prior probability of the image to be processed that is as large as the image template.
- Fig. 1 is a flowchart of a camera positioning method according to an exemplary embodiment of the present disclosure
- Fig. 2 is a flowchart of step 110 according to an exemplary embodiment of the present disclosure
- Fig. 3 is a schematic diagram showing an image template according to an exemplary embodiment of the present disclosure
- Fig. 4 is a flowchart showing step 120 according to an exemplary embodiment of the present disclosure.
- Fig. 5 is a flowchart of a camera positioning method according to another exemplary embodiment of the present disclosure.
- Fig. 6 is a flowchart of step 150 according to an exemplary embodiment of the present disclosure.
- Fig. 7 is a schematic diagram showing multiple absolute poses according to an exemplary embodiment of the present disclosure.
- Fig. 8 is a schematic diagram showing a process of determining and correcting a pose according to an exemplary embodiment of the present disclosure
- Fig. 9 is a schematic diagram showing an optimized pose graph according to an exemplary embodiment of the present disclosure.
- Fig. 10 is a flowchart of a camera positioning method according to another exemplary embodiment of the present disclosure.
- Fig. 11 is a flowchart showing step 230 according to an exemplary embodiment of the present disclosure.
- 12A to 12B are schematic diagrams showing a self-attention mechanism according to an exemplary embodiment of the present disclosure
- Fig. 13A is a schematic diagram showing an image to be processed according to an exemplary embodiment of the present disclosure
- Fig. 13B is a schematic diagram showing a feature extraction image after weight value adjustment according to an exemplary embodiment of the present disclosure
- Fig. 14 is a flowchart of a camera positioning method according to another exemplary embodiment of the present disclosure.
- Fig. 15 is a frame diagram of a target neural network according to an exemplary embodiment of the present disclosure.
- Fig. 16 is a block diagram showing a camera positioning device according to an exemplary embodiment of the present disclosure.
- Fig. 17 is a block diagram showing an obtaining module according to an exemplary embodiment of the present disclosure.
- Fig. 18 is a block diagram showing an execution module according to an exemplary embodiment of the present disclosure.
- Fig. 19 is a block diagram showing a positioning module according to an exemplary embodiment of the present disclosure.
- Fig. 20 is a block diagram showing a camera positioning device according to another exemplary embodiment of the present disclosure.
- Fig. 21 is a block diagram showing a second determining module according to an exemplary embodiment of the present disclosure.
- Fig. 22 is a block diagram showing a camera positioning device according to another exemplary embodiment of the present disclosure.
- Fig. 23 is a block diagram showing an obtaining module according to an exemplary embodiment of the present disclosure.
- Fig. 24 is a block diagram showing an execution module according to an exemplary embodiment of the present disclosure.
- Fig. 25 is a block diagram showing a positioning module according to an exemplary embodiment of the present disclosure.
- Fig. 26 is a block diagram showing a camera positioning device according to another exemplary embodiment of the present disclosure.
- Fig. 27 is a schematic structural diagram of an electronic device according to an exemplary embodiment of the present disclosure.
- first, second, third, etc. may be used in this disclosure to describe various information, the information should not be limited to these terms. These terms are only used to distinguish the same type of information from each other.
- first information may also be referred to as second information, and similarly, the second information may also be referred to as first information.
- word “if” as used herein can be interpreted as "when” or “when” or “in response to a certainty”.
- the embodiments of the present disclosure provide a camera positioning method, which can discard parts of the image to be processed that are as large as the image template according to the prior probability of a movable object appearing at each of the multiple pixels included in the image template
- the pixel points are used to obtain the target image, and then the absolute pose of the camera is determined according to the target image, which reduces the influence of the movement of objects in the scene where the camera collects the image on the camera positioning result, and improves the accuracy and precision of the camera positioning.
- the camera positioning method provided by the embodiments of the present disclosure can be applied to a movable machine and equipment to position a camera provided on the movable machine and equipment.
- Movable machinery and equipment include, but are not limited to, drones, unmanned vehicles, and robots with cameras.
- the accuracy of camera positioning can improve the accuracy of mobile machinery and equipment when performing various tasks. For example, according to the image of the forward environment of the vehicle collected by the camera installed on the unmanned vehicle, the current location information of the camera can be determined, and the current location information of the vehicle can be located according to the location information of the camera, so that the unmanned vehicle can be The vehicle performs at least one intelligent driving control such as path planning, trajectory tracking, and collision warning.
- the camera positioning method provided by the embodiment of the present disclosure may include the following steps 110-130:
- step 110 the prior probability of a movable object appearing at each of the multiple pixels included in the image template is obtained.
- the image template may include a template corresponding to the current scene and used to record the prior probability of a movable object appearing at each of multiple pixels on an image as large as the image template.
- Movable objects include, but are not limited to, various objects that can move on their own or under control, such as buses, cars, people, bicycles, trucks, motorcycles, animals, etc.
- the prior probability refers to the probability that each pixel on the image is a movable object obtained by analyzing the image that is the same or similar to the current scene collected in the past.
- the prior probability corresponding to a certain pixel is high, it means that there is a high possibility of a movable object at that pixel in the image collected for the scene; conversely, if the prior probability corresponding to a certain pixel is low, it means that The possibility of a movable object at this pixel point in the image collected by the scene is low.
- the image template can reflect the a priori possibility of movable objects appearing at different pixels in the collected image.
- the probability of a movable object appearing at each pixel on each image in the above image collection can be analyzed for a collection of images collected from a scene that is the same or similar to the current scene, and this probability can be used as each image template corresponding to the current scene.
- the prior probability that a movable object appears at each pixel can be analyzed for a collection of images collected from a scene that is the same or similar to the current scene, and this probability can be used as each image template corresponding to the current scene.
- the prior probability that a movable object appears at each pixel can be used as each image template corresponding to the current scene.
- the image collection collected in the same or similar scene as the current scene can include the main street of the city. At least one image.
- step 120 an operation of discarding some pixels is performed on an image to be processed that is as large as the image template according to the prior probability to obtain a target image.
- the image to be processed may be at least one image collected by a camera provided on the movable machine equipment during the movement of the movable machine equipment.
- the mobile machinery and equipment can perform the discarding part on at least one image that is as large as the image template collected by the camera set on the mobile machinery and equipment according to the prior probability corresponding to each pixel on the image template corresponding to the current scene The operation of the pixel points to obtain the target image.
- the operation of discarding part of the pixels includes, but is not limited to, discarding all or all pixels whose a priori probability sampling value is greater than a preset value on at least one image collected by the camera and the size of the image template. Partially discarded randomly.
- step 130 the absolute pose of the camera collecting the image to be processed in the world coordinate system is determined according to the target image.
- the mobile machine equipment can determine the absolute pose of the camera in the world coordinate system according to the target image through regression loss function.
- the regression loss function can be a mean square error loss function (such as L2 loss function), average absolute error (such as L1 loss function), smooth average absolute error loss function (such as Huber loss function), logarithmic hyperbolic cosine loss function or Quantile loss function, etc.
- the movable machinery and equipment can be combined with the prior probability that a movable object appears at each of the multiple pixels on the image template corresponding to the current scene. At least one of the collected images is discarded with some pixels to obtain the target image, and the target image is used to determine the absolute pose of the camera, which can effectively reduce the negative impact of the movement of objects in the current scene on the camera positioning and improve the camera positioning Accuracy and precision.
- a camera set on a movable machine equipment its pose may be changed due to factors such as the movement of the movable machine equipment and/or the position adjustment of the camera, so that the camera needs to be positioned.
- the inventor of the present disclosure found that if there is movement of an object in the field of view of the image captured by the camera, the movement of the object will cause poor imaging quality of the corresponding part of the image captured by the camera, such as image blur, jitter, etc. Poor quality parts will affect the quality of the overall features of the captured image, and further affect the accuracy and precision of camera positioning based on the overall features of the image. However, some immobile or fixed objects in the captured image are actually useful for camera positioning.
- the embodiments of the present disclosure determine the probability of a movable object appearing at each pixel in the captured image (that is, the prior probability) by combining prior knowledge, and perform partial pixel discarding on the captured image based on the determined probability , Such as discarding some of the pixels with a higher a priori probability of moving objects, which can reduce the negative impact of these pixels on the overall quality of the image, which is beneficial to improve the overall quality of the image after the local pixel is discarded.
- the accuracy of positioning is the probability of a movable object appearing at each pixel in the captured image (that is, the prior probability) by combining prior knowledge, and perform partial pixel discarding on the captured image based on the determined probability .
- step 110 may be performed by an electronic device, which may be a mobile machine device, or an electronic device for training a neural network, such as a cloud platform, which is not limited in the present disclosure. As shown in Figure 2, step 110 may include steps 111-113:
- step 111 pixel-level semantic segmentation is performed on each image in a predetermined image set associated with the current scene.
- the predetermined image set associated with the current scene includes multiple pictures collected in the same or similar scene as the current scene.
- the electronic device can obtain the pixel-level semantic segmentation result of each image by searching for the content existing on each image in the predetermined image set.
- the predetermined image set associated with the current scene may include images m1, m2,...mN as shown in FIG. 3.
- step 112 the first pixel belonging to the movable object and the second pixel belonging to the background in each image are determined according to the result of pixel-level semantic segmentation.
- the background is an immovable object in the image, for example, other objects in the image that are not determined to be movable objects, such as sky, buildings, trees, roads, etc.
- step 113 based on the statistical distribution of the first pixel and the second pixel on each image in the predetermined image set, it is determined that each of the multiple pixels included in the image template that is the same size as the image in the predetermined image set The prior probability that a movable object appears at each pixel.
- the electronic device obtains a statistical distribution corresponding to the current scene based on the statistical distribution of the first pixel of the movable object and the second pixel of the background in each image in the predetermined image set associated with the current scene
- An image template such as the image template M in FIG. 3, is used to record the prior probability that a movable object appears at each pixel in an image that is as large as the image template collected in the current scene.
- the prior probability of a movable object at each pixel recorded on the image template is a statistical distribution range, rather than a fixed value.
- the subsequent operation of discarding some pixels of an image to be processed that is as large as the image template is performed according to the prior probability, different pixels can be discarded according to the statistical distribution range of the prior probability each time to obtain different targets.
- image determining the absolute pose of the camera based on multiple different target images can obtain better camera positioning results, especially in large-scale urban traffic scenes.
- the prior probability that a movable object appears at each pixel included in the image template may conform to a Gaussian distribution, as shown in formula 1:
- i represents the pixel point of the i-th row on the image template
- j represents the pixel point of the j-th column on the image template
- (i,j) corresponds to the pixel coordinates
- the mathematical expectation of the pixel point (i,j) is ⁇ (i ,j)
- N is the number of pixels
- the variance of the pixel (i,j) is ⁇ 2 (i,j)
- ⁇ 2 (i,j) ⁇ (i,j)(1- ⁇ (i,j))
- P(M(i,j)) is the prior probability of pixel (i,j).
- step 120 may include:
- step 121 a priori probability corresponding to at least some pixels included in the image to be processed is sampled.
- the distribution of the prior probability that a movable object appears at each pixel on each to-be-processed image satisfies a Gaussian distribution.
- the mobile machine and equipment can sample the prior probability corresponding to at least some of the pixels included in the to-be-processed image to obtain the image on the to-be-processed image after this sampling.
- the sampling value of the prior probability corresponding to at least some pixels.
- step 122 pixel points whose a priori probability sampling value is greater than a preset threshold are removed from the image to be processed to obtain a target image corresponding to this sampling.
- the mobile machine equipment can remove all pixels with a priori probability that the sampling value is greater than a preset threshold for the to-be-processed image in the above-mentioned manner, or randomly remove part of the priori probability For pixel points whose sampling value is greater than the preset threshold value, the target image corresponding to the current sampling of the image to be processed is obtained.
- the prior probability of the same pixel on the same image to be processed can be The sampling values are different, so that there is at least one different pixel point between the multiple target images obtained after the operation of discarding some pixels.
- the sampling value of the prior probability corresponding to pixel 1 on image 1 to be processed is P 1
- the sampling of the prior probability corresponding to pixel 1 on image 1 to be processed The value is P 2 and the preset threshold is T. Among them, P 1 ⁇ T ⁇ P 2 . Then the target image obtained after the first sampling retains pixel 1, and the target image obtained after the second sampling needs to remove pixel 1.
- the mobile equipment can sample the prior probabilities corresponding to the pixels on the same image to be processed multiple times, and accordingly obtain multiple different target images for camera positioning, which is beneficial to guarantee the final result The accuracy of camera positioning.
- step 130 may include: inputting the to-be-processed image into a target neural network to obtain the absolute pose of the camera in the world coordinate system.
- the mobile machine equipment can input the image to be processed into the target neural network, and the target neural network directly outputs the absolute pose of the camera that collects the image to be processed in the world coordinate system.
- the movable machine equipment discards at least some pixels on the image to be processed with a prior probability greater than a preset value according to the prior probability that each pixel on the image template belongs to the movable object, thereby improving the camera positioning Accuracy.
- the image to be processed includes k frames of images (k is an integer greater than or equal to 2) that are acquired by the camera in time sequence, that is, time sequence, as shown in FIG. 5,
- the method also includes steps 140-150:
- step 140 the relative pose of the camera when shooting the k frames of images is determined according to the k frames of images.
- the movable machine equipment can use the visual odometry method to determine the relative pose of the camera when acquiring the k-th frame image with respect to the acquisition of the k-1 frame image.
- step 150 the corrected pose of the camera is determined according to the relative pose and absolute pose of the camera.
- the mobile machine equipment can use the camera's absolute pose in the world coordinate system when acquiring the first frame image (also referred to as the first frame image) in the sequence of k frames of images as a reference , According to the relative pose and absolute pose of the camera when the second frame image adjacent to the first frame image is collected, the corrected pose of the camera is determined.
- the first frame image also referred to as the first frame image
- the corrected pose of the camera is determined.
- Subsequent movable machinery and equipment can adjust the pose of the camera according to the corrected pose, thereby reducing the impact of the movement of objects in the scene on the positioning of the camera, which can help ensure the accuracy of the movable machinery and equipment in performing various tasks.
- step 150 may specifically include steps 151-153:
- step 151 the deterministic probability of the absolute pose is determined.
- the deterministic probability is an evaluation of the accuracy of the result of the absolute pose. If the probability of certainty is higher, the result of the absolute pose is more accurate, otherwise the result of the absolute pose is less accurate.
- the movable machinery and equipment can adopt a random sampling method, such as Monte Carlo method, to sample the prior probabilities corresponding to the k-frame images with sequential nature collected by the camera to obtain the sampling result of multiple sampling.
- k is an integer greater than or equal to 2.
- the current image can be sampled multiple times based on the prior probability of each pixel included in the image template M, and multiple absolute positions corresponding to the current image can be determined based on the target image corresponding to each sample. posture.
- the deterministic probability of the absolute pose corresponding to the current image is determined according to the multiple absolute poses corresponding to the current image. For example, if the difference between the absolute poses corresponding to the current image is large, it can be determined that the absolute pose corresponding to the current image has a low probability of certainty. Otherwise, the absolute pose corresponding to the current image is determined to be certain The probability is higher.
- step 152 the first weight of the relative pose and the second weight of the absolute pose are determined according to the deterministic probability of the absolute pose.
- the movable machine equipment can determine the first weight of the relative pose corresponding to each frame of image according to the deterministic probability of the absolute pose corresponding to each frame of image And the second weight of the absolute pose corresponding to each frame of image.
- the second weight of the absolute pose corresponding to the current image can be increased; if the absolute pose corresponding to the current image has a low certainty probability, it can be increased The first weight of the relative pose corresponding to the current image.
- step 153 the corrected pose of the camera is determined according to the relative pose, the first weight, the absolute pose, and the second weight.
- the sliding window is adopted to move sequentially, and the relative position corresponding to the second frame of image
- the pose, the first weight, the absolute pose and the second weight determine the corrected pose of the second frame of image relative to the first frame of image.
- the weight of the relative pose can be increased, and if the absolute pose is more accurate, the weight of the absolute pose can be increased. In this way, by making the relative pose and the absolute pose each have different weights to determine the corrected pose, the corrected pose can be made more accurate, and the camera positioning can be more accurate.
- the optimized pose map can be shown in Figure 9.
- the triangle in Figure 9 represents the absolute pose when the camera collects each frame of image, with The arrow line represents the relative pose, and the circle represents the sliding window.
- the corrected absolute pose and relative pose in Fig. 9 correspond to the absolute pose and relative pose in Fig. 8 from the upper left corner to the lower right corner in sequence according to the arrow direction.
- the pose determined by the VO (Visual Odometry) method may be used as the relative pose corresponding to the image.
- the VO method is to determine the position and posture of the camera by analyzing the above k frames of images. Estimating the movement of the camera between adjacent frames by performing feature matching on k frames of images, so as to obtain the relative pose of the camera when the next frame is collected compared to the previous frame.
- the absolute pose and relative pose are combined to perform pose correction, which further improves the accuracy of camera positioning.
- the camera positioning method provided in the present disclosure can also be applied to electronic devices that train neural networks, such as cloud platforms, neural network training platforms, and so on.
- the electronic device uses this method to train the neural network to obtain the target neural network. After the image is subsequently input to the target neural network, the absolute pose of the camera that collected the image in the world coordinate system can be obtained.
- the camera positioning method provided by the embodiment of the present disclosure may include the following steps 210-230:
- step 210 the prior probability that a movable object appears at each of the multiple pixels included in the image template is obtained.
- the electronic device can analyze the probability of a movable object appearing at each pixel of each image based on each of the above images, and use this probability as the occurrence of a movable object at each pixel on an image template that is the same size as each image. The prior probability.
- step 220 according to the prior probability, an operation of discarding part of pixels is performed on an image to be processed that is as large as the image template to obtain a target image.
- the image to be processed may be at least one sample image, and the electronic device may perform the operation of discarding some pixels on the at least one sample image according to the prior probability corresponding to each pixel on the image template, so as to obtain the target image.
- the operation of discarding some pixels includes but is not limited to the operation of discarding all pixels or randomly partially discarding pixels on at least one sample image whose a priori probability sampling value is greater than a preset value.
- step 230 the absolute pose of the camera that collects the image to be processed in the world coordinate system is determined according to the target image.
- the electronic device can determine the absolute pose of the camera that collects at least one sample image in the world coordinate system through the regression loss function according to the obtained target image.
- the regression loss function can be a mean square error loss function (such as L2 loss function), average absolute error (such as L1 loss function), smooth average absolute error loss function (such as Huber loss function), log hyperbolic cosine loss function, and Quantile loss function, etc.
- step 210 may be performed by an electronic device that trains a neural network, and the execution process is the same as the execution process of step 110 in FIG. 2, and will not be repeated here.
- step 220 may be performed by an electronic device that trains a neural network, and the execution process is the same as that of step 120 in FIG. 4, and will not be repeated here.
- step 230 may be performed by an electronic device that trains a neural network.
- step 230 may include steps 231-233:
- step 231 the feature parameters in the target image are extracted through a neural network to obtain a feature extraction image.
- the neural network can extract feature parameters of each target image from at least one target image, thereby obtaining a feature extraction image corresponding to each target image.
- step 232 on the preset spatial dimension and/or preset channel dimension of the neural network, the weight value corresponding to the second pixel point belonging to the background in the feature extraction image is increased.
- the neural network can increase the weight value of the second pixel point belonging to the background in the feature extraction image in at least one of the preset space dimension and the preset channel dimension through a self-attention mechanism.
- the neural network transforms a certain feature extraction image of H (height) ⁇ W (width) ⁇ C (channel) using a spatial self-attention mechanism to obtain an image H ⁇ W ⁇ 1 on the same channel.
- the neural network transforms a certain feature extraction image of H ⁇ W ⁇ C using the channel self-attention mechanism to obtain an image 1 ⁇ 1 ⁇ C with the same height and width.
- the neural network ignores the information of the first pixel of the movable object as much as possible, and pays more attention to the information of the second pixel of the background.
- the image shown in FIG. 13B is obtained.
- the gray values of pixels in the image shown in FIG. 13B are higher than the gray values of pixels in other parts of the image shown in FIG. 13B.
- the pixels circled by the dashed box belong to the movable object automobile, and each of the image templates that are as large as the image shown in FIG. 13A can be obtained through the previous step 210.
- the prior probability that a movable object appears at the pixel point and then step 220 discards all or part of the pixel point in the image shown in FIG. 13A whose sampling value of the prior probability is greater than the preset threshold.
- the weight values belonging to immovable objects are increased in two dimensions, so that the neural network pays more attention to traffic signs, electric poles and other immovable or low-moving objects, reducing the location where the camera collects images.
- the effect of the movement of objects in the scene on the result of the positioning of the camera on the movable machine equipment improves the accuracy and precision of the positioning of the camera by the neural network, and improves the robustness of the positioning detection result.
- step 233 the feature extraction image adjusted by the weight value is analyzed by the neural network to obtain the absolute pose of the camera that collects the image to be processed in the world coordinate system.
- the neural network can analyze the feature extraction image after the weight value adjustment through regression loss function, such as the mean square error function, the absolute value error function, etc., to obtain the camera that collects at least one sample image in the world coordinate system The absolute pose of the next.
- regression loss function such as the mean square error function, the absolute value error function, etc.
- the above-mentioned camera positioning method further includes step 240:
- step 240 according to the difference between the absolute pose and the predetermined true value of the camera's pose of the image to be processed, the network parameters of the neural network are adjusted to obtain the target neural network by training.
- this step may be performed by an electronic device that trains a neural network.
- the camera acquires at least one sample image that is as large as the image template, the true value of the pose is known.
- the electronic device can use the neural network output to collect at least one sample image of the camera's absolute pose and data in the world coordinate system. Knowing the difference in the true value of the pose, adjust the network parameters of the neural network to minimize the loss function of the neural network, and finally train the desired target neural network.
- the embodiments of the present disclosure are based on the above-mentioned camera positioning method, and also provide a framework diagram of a target neural network.
- a target neural network For example, as shown in FIG. 15, it includes Probabilistic Dropout Module (Partial Pixel Dropout Module), Feature Ectractor Module (feature extraction module), Self-attention Module (self-attention module) and Regressor Module (regression module).
- At least one sample image may be used as the input value of the partial pixel discarding module, and the partial pixel discarding module may be composed of at least five sub-networks connected in sequence.
- Each sub-network can be implemented separately by using network units set in a preset order, such as a convolutional layer, a Relu layer, and a pooling layer.
- the first sub-network can perform pixel-level semantic segmentation on each image in at least one sample image; the second sub-network can determine the first part of each sample image belonging to the movable object according to the result of pixel-level semantic segmentation. A pixel and a second pixel belonging to the background; the third sub-network can determine an image template as large as the sample image based on the statistical distribution of the first pixel and the second pixel in each sample image.
- the fourth sub-network may sample the prior probability corresponding to at least some of the pixels included in at least one sample image to obtain the original The sampling result of the sub-sampling; the fifth sub-network can remove the pixel points with the sampling value of the prior probability greater than the preset threshold T in at least one sample image according to the sampling result of this time to obtain the target image.
- the feature extraction module can be designed by stacking network units set in a preset order such as convolutional layer, Relu layer, pooling layer, etc. according to the preset structure, and extract the feature parameters in the target image obtained by Probabilistic Dropout Module to obtain feature extraction image.
- the self-attention module can also be composed of at least two separate fifth and sixth sub-networks.
- Each sub-network includes a convolutional layer, a Relu layer, a pooling layer and other network units set in a preset order.
- the fifth The sub-network can focus on the preset spatial dimension
- the sixth sub-network can focus on the preset channel dimension. After passing through the above two sub-networks, the weight value of the second pixel that belongs to the background in the feature extraction image can be adjusted.
- the embodiment of the present disclosure does not limit the sequence of the fifth sub-network and the sixth sub-network.
- the regression module may include a seventh sub-network, the seventh sub-network may include a convolutional layer, a Relu layer, a pooling layer and other network units set in a preset order, and the seventh sub-network takes the image output by the self-attention module as input value , Taking the known pose of the camera that collects at least one sample image as the output value, and the seventh sub-network corresponds to a regression loss function.
- the regression loss function can include mean square error loss function (such as L2 loss function), average absolute error (such as L1 loss function), smooth average absolute error loss function (such as Huber loss function), log hyperbolic cosine loss function, and score Bit loss function, etc.
- the target neural network finally obtained reduces the focus on the movable objects on the sample image, and pays more attention to the background pixels on the sample image, that is, the information of immobile or fixed objects.
- the impact of the pixel points corresponding to the object on the overall imaging quality of the image improves the robustness of the target neural network.
- the present disclosure also provides an embodiment of a camera positioning device.
- the embodiments of the present disclosure also provide a camera positioning device, which can be applied to movable machinery and equipment. Since the movable electronic equipment will move, the pose of the camera set on the movable machinery and equipment will change accordingly. The high accuracy of camera positioning can improve the accuracy of mobile machinery and equipment when performing various tasks.
- FIG. 16 is a block diagram of a camera positioning device according to an exemplary embodiment of the present disclosure.
- the device includes: an acquisition module 310, configured to acquire each of the multiple pixels included in the image template The prior probability of a movable object appearing at the location; the execution module 320 is configured to perform the operation of discarding some pixels according to the prior probability for the image to be processed as large as the image template to obtain the target image; the positioning module 330, It is used to determine the absolute pose of the camera that collects the image to be processed in the world coordinate system according to the target image.
- the acquisition module 310 includes: a segmentation sub-module 311, configured to perform pixel-level semantic segmentation on each image in a predetermined image set; and a first determination sub-module 312, configured to According to the result of pixel-level semantic segmentation, determine the first pixel that belongs to the movable object and the second pixel that belongs to the background in each image; the second determination sub-module 313 is used to determine based on the The statistical distribution of the first pixel point and the second pixel point is determined, and all the pixels where the movable object appears at each of the multiple pixel points included in the image template as large as the image in the predetermined image set are determined. State the prior probability.
- the execution module 320 includes: a sampling sub-module 321, configured to sample the prior probability corresponding to at least some of the pixels included in the image to be processed; and execute;
- the sub-module 322 is configured to remove the pixel points whose a priori probability sampling value is greater than a preset threshold on the image to be processed to obtain the target image.
- the positioning module 330 includes a second positioning sub-module 331 for inputting the image to be processed into the target neural network, and obtaining the camera of the image to be processed The absolute pose in the world coordinate system.
- the image to be processed includes at least two frames of images with time sequence captured by the camera; for example, as shown in FIG. 20, the device further includes: a first determining module 340, configured to Two frames of images determine the relative pose of the camera when shooting the at least two frames of images; the second determining module 350 is configured to determine the correction of the camera according to the relative pose of the camera and the absolute pose Posture.
- a first determining module 340 configured to Two frames of images determine the relative pose of the camera when shooting the at least two frames of images
- the second determining module 350 is configured to determine the correction of the camera according to the relative pose of the camera and the absolute pose Posture.
- the second determining module 350 further includes: a third determining sub-module 351, configured to determine the deterministic probability of the absolute pose; and a fourth determining sub-module 352, configured to In order to determine the first weight of the relative pose and the second weight of the absolute pose according to the deterministic probability; a fifth determination sub-module 353 is configured to determine the relative pose, the first weight, The absolute pose and the second weight determine the corrected pose of the camera.
- the present disclosure also provides a camera positioning device, which can be applied to an electronic device, and the electronic device can train a neural network to obtain a target neural network. After the image is subsequently input to the target neural network, the absolute pose of the camera that collected the image in the world coordinate system can be obtained.
- FIG. 22 is a block diagram of a camera positioning device according to an exemplary embodiment of the present disclosure.
- the device includes: an acquisition module 410, configured to acquire each of the multiple pixels included in the image template The prior probability of a movable object appearing at the location; the execution module 420 is configured to perform the operation of discarding some pixels according to the prior probability for the image to be processed as large as the image template to obtain the target image; the positioning module 430, It is used to determine the absolute pose of the camera that collects the image to be processed in the world coordinate system according to the target image.
- the acquisition module 410 includes: a segmentation sub-module 411, configured to perform pixel-level semantic segmentation on each image in a predetermined image set; and a first determining sub-module 412, configured to According to the result of pixel-level semantic segmentation, the first pixel that belongs to the movable object and the second pixel that belongs to the background in each image are determined; the second determination sub-module 413 is used to determine the The statistical distribution of the first pixel point and the second pixel point is determined, and all the pixels where the movable object appears at each of the multiple pixels included in the image template that is as large as the image in the predetermined image set are determined. State the prior probability.
- the execution module 420 includes: a sampling sub-module 421, configured to sample the prior probability corresponding to at least some pixels included in the image to be processed; execute;
- the sub-module 422 is configured to remove pixels with a priori probability sampling value greater than a preset threshold on the image to be processed to obtain the target image.
- the positioning module 430 includes: a first processing sub-module 431 for extracting feature parameters in the target image via a neural network to obtain a feature extraction image; a second processing sub-module The module 432 is configured to increase the weight value corresponding to the second pixel point belonging to the background in the feature extraction image in the preset space dimension and/or the preset channel dimension of the neural network; the first positioning sub-module 433 , Used to analyze the feature extraction image after the weight value is adjusted by the neural network, and obtain the absolute pose of the camera that collects the image to be processed in the world coordinate system.
- the device further includes: a training module 440 configured to determine the true value of the camera's pose and pose according to the absolute pose and the predetermined image to be processed. Differences, adjust the network parameters of the neural network, and train the target neural network.
- a training module 440 configured to determine the true value of the camera's pose and pose according to the absolute pose and the predetermined image to be processed. Differences, adjust the network parameters of the neural network, and train the target neural network.
- the relevant part can refer to the part of the description of the method embodiment.
- the device embodiments described above are merely illustrative.
- the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place. , Or it can be distributed to multiple network units.
- Some or all of the modules can be selected according to actual needs to achieve the objectives of the solutions of the present disclosure. Those of ordinary skill in the art can understand and implement it without creative work.
- the embodiment of the present disclosure also provides a computer-readable storage medium, the storage medium stores a computer program, and the computer program is used to execute any of the above-mentioned camera positioning methods.
- the embodiment of the present disclosure also provides a camera positioning device, the device includes: a processor; a memory for storing executable instructions of the processor; wherein the processor is used for calling the executable instructions stored in the memory to implement any one of the aforementioned cameras Positioning method.
- the camera positioning device provided in the embodiments of the present disclosure can implement the method provided in any of the foregoing embodiments.
- the camera positioning device can discard some pixels in the image to be processed that are as large as the image template according to the prior probability of a movable object appearing at each of the multiple pixels included in the image template, and then according to the obtained
- the target image is used to determine the absolute pose of the camera, which reduces the influence of the movement of the object in the scene where the camera collects the image on the result of the positioning of the camera on the movable machinery and equipment, and improves the accuracy of the camera positioning.
- the camera positioning device provided by the embodiments of the present disclosure can be applied to movable machinery and equipment to locate cameras provided on the movable machinery and equipment. Since the movable machinery and equipment will move, the pose of the camera set on the equipment will change accordingly.
- the accuracy of camera positioning can improve the accuracy of mobile machinery and equipment when performing various tasks. For example, according to the image of the forward environment of the vehicle collected by the camera installed on the unmanned vehicle, the current location information of the camera can be determined, and the current location information of the vehicle can be located according to the location information of the camera, so that the unmanned vehicle can be The vehicle performs at least one intelligent driving control such as path planning, trajectory tracking, and collision warning.
- the camera positioning device provided by the present disclosure can also be used on electronic devices for training neural networks, such as cloud platforms, neural network training platforms, and the like.
- the electronic device uses this method to train the neural network to obtain the target neural network. After the image is subsequently input to the target neural network, the absolute pose of the camera that collected the image in the world coordinate system can be obtained.
- Fig. 27 is a schematic structural diagram of an electronic device 2700 according to an exemplary embodiment.
- the electronic device 2700 includes movable machinery and a cloud platform for training neural networks.
- the electronic device 2700 includes a processing component 2722, which further includes one or more processors, and a memory resource represented by a memory 2732 for storing instructions executable by the processing component 2722, such as application programs.
- the application program stored in the memory 2732 may include at least one module, and each module corresponds to a set of instructions.
- the processing component 2722 is used to execute instructions to execute any of the aforementioned camera positioning methods.
- the electronic device 2700 may further include a power component 2726 for performing power management of the electronic device 2700, a wired or wireless network interface 2750 for connecting the electronic device 2700 to a network, and an input output (I/O) interface 2758.
- the electronic device 2700 can operate based on an operating system stored in the memory 2732, such as Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeB SDTM or the like.
- the electronic device 2700 is a movable machine device
- the electronic device 2700 further includes a camera for capturing images.
- the electronic device 2700 is a cloud platform for training a neural network
- the electronic device can communicate with a mobile machine device through the input and output interface 2758.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (20)
- 一种相机定位方法,包括:A camera positioning method includes:获取图像模板包括的多个像素点中每个像素点处出现可移动物体的先验概率;Obtain the prior probability of a movable object appearing at each of the multiple pixels included in the image template;根据所述先验概率针对与所述图像模板等大的待处理图像执行丢弃部分像素点的操作,得到目标图像;Performing an operation of discarding some pixels for an image to be processed that is as large as the image template according to the prior probability to obtain a target image;根据所述目标图像确定采集所述待处理图像的相机在世界坐标系下的绝对位姿。The absolute pose of the camera that collects the image to be processed in the world coordinate system is determined according to the target image.
- 根据权利要求1所述的方法,其特征在于,获取所述图像模板包括的多个像素点中每个像素点处出现可移动物体的先验概率,包括:The method according to claim 1, wherein obtaining the prior probability of a movable object at each of the plurality of pixels included in the image template comprises:对预定图像集合中的每张图像进行像素级语义分割;Perform pixel-level semantic segmentation on each image in the predetermined image set;根据像素级语义分割的结果确定所述每张图像中属于可移动物体的第一像素点和属于背景的第二像素点;Determining, according to the result of pixel-level semantic segmentation, the first pixel that belongs to the movable object and the second pixel that belongs to the background in each image;基于所述每张图像中所述第一像素点和所述第二像素点的统计分布,确定与所述预定图像集合中的图像等大的图像模板包括的多个像素点中每个像素点处出现所述可移动物体的所述先验概率。Based on the statistical distribution of the first pixel point and the second pixel point in each image, determine each pixel point in a plurality of pixels included in an image template that is as large as an image in the predetermined image set The prior probability that the movable object appears at.
- 根据权利要求1或2所述的方法,其特征在于,根据所述先验概率对所述待处理图像执行丢弃部分像素点的操作,得到目标图像,包括:The method according to claim 1 or 2, wherein the step of discarding some pixels of the image to be processed according to the prior probability to obtain the target image comprises:对所述待处理图像所包括的至少部分像素点对应的先验概率进行采样;Sampling the prior probabilities corresponding to at least some of the pixels included in the image to be processed;在所述待处理图像上去除先验概率的采样值大于预设阈值的像素点,得到所述目标图像。Remove the pixel points with a priori probability sampling value greater than a preset threshold from the image to be processed to obtain the target image.
- 根据权利要求3所述的方法,其特征在于,采样次数为多次时,执行丢弃部分像素点的操作之后得到的多个目标图像两两之间存在至少一个不同的像素点。The method according to claim 3, wherein when the number of sampling times is multiple times, there is at least one different pixel point between the multiple target images obtained after the operation of discarding some pixels.
- 根据权利要求1-4任一项所述的方法,其特征在于,根据所述目标图像确定采集所述待处理图像的相机在世界坐标系下的绝对位姿,包括:The method according to any one of claims 1 to 4, wherein determining, according to the target image, the absolute pose of the camera that collects the image to be processed in the world coordinate system comprises:经神经网络提取所述目标图像中的特征参数,得到特征提取图像;Extracting feature parameters in the target image through a neural network to obtain a feature extraction image;在所述神经网络的预设空间维度和/或预设通道维度上,增加所述特征提取图像中属于背景的第二像素点所对应的权重值;Adding a weight value corresponding to a second pixel that belongs to the background in the feature extraction image on the preset space dimension and/or the preset channel dimension of the neural network;经所述神经网络对权重值调整后的特征提取图像进行分析,得到采集所述待处理图像的相机在世界坐标系下的所述绝对位姿。The neural network analyzes the feature extraction image after the weight value is adjusted to obtain the absolute pose of the camera that collects the image to be processed in the world coordinate system.
- 根据权利要求5所述的方法,其特征在于,经所述神经网络对权重值调整后的特征提取图像进行分析,得到采集所述待处理图像的相机在世界坐标系下的所述绝对位姿之后,所述方法还包括:The method according to claim 5, wherein the feature extraction image after the weight value adjustment is analyzed by the neural network to obtain the absolute pose of the camera collecting the image to be processed in the world coordinate system After that, the method further includes:根据所述绝对位姿和预先确定的采集所述待处理图像的所述相机的位姿真值的差异,调整所述神经网络的网络参数,训练得到目标神经网络。Adjust the network parameters of the neural network according to the difference between the absolute pose and the predetermined true value of the camera that collects the image to be processed to obtain a target neural network through training.
- 根据权利要求6所述的方法,其特征在于,根据所述目标图像确定采集所述待处理图像的相机在世界坐标系下的绝对位姿,包括:The method according to claim 6, wherein determining the absolute pose of the camera that collects the image to be processed in the world coordinate system according to the target image comprises:将所述待处理图像输入所述目标神经网络,得到采集所述待处理图像的相机在世界坐标系下的所述绝对位姿。The image to be processed is input into the target neural network to obtain the absolute pose of the camera that collects the image to be processed in the world coordinate system.
- 根据权利要求1-7任一项所述的方法,其特征在于,所述待处理图像包括所述相机采集的具有时序性的至少两帧图像;The method according to any one of claims 1-7, wherein the image to be processed comprises at least two frames of images with time series collected by the camera;根据所述目标图像确定采集所述待处理图像的相机在世界坐标系下的绝对位姿之后,所述方法还包括:After determining the absolute pose of the camera that collects the image to be processed in the world coordinate system according to the target image, the method further includes:根据所述至少两帧图像确定所述相机在拍摄所述至少两帧图像时的相对位姿;Determining, according to the at least two frames of images, the relative pose of the camera when shooting the at least two frames of images;根据所述相机的相对位姿和所述绝对位姿,确定所述相机的修正位姿。Determine the corrected pose of the camera according to the relative pose of the camera and the absolute pose.
- 根据权利要求8所述的方法,其特征在于,根据所述相机的相对位姿和所述绝对位姿,确定所述相机的修正位姿包括:The method according to claim 8, wherein determining the corrected pose of the camera according to the relative pose of the camera and the absolute pose comprises:确定所述绝对位姿的确定性概率;Determining the certainty probability of the absolute pose;根据所述确定性概率确定所述相对位姿的第一权重和所述绝对位姿的第二权重;Determining the first weight of the relative pose and the second weight of the absolute pose according to the certainty probability;根据所述相对位姿、所述第一权重、所述绝对位姿和所述第二权重,确定所述相机的修正位姿。Determine the corrected pose of the camera according to the relative pose, the first weight, the absolute pose, and the second weight.
- 一种相机定位装置,包括:A camera positioning device includes:获取模块,用于获取图像模板包括的多个像素点中每个像素点处出现可移动物体的先验概率;An acquisition module for acquiring the prior probability of a movable object appearing at each of the multiple pixels included in the image template;执行模块,用于根据所述先验概率针对与所述图像模板等大的待处理图像执行丢弃部分像素点的操作,得到目标图像;The execution module is configured to perform an operation of discarding some pixels for an image to be processed that is as large as the image template according to the prior probability to obtain a target image;定位模块,用于根据所述目标图像确定采集所述待处理图像的相机在世界坐标系下的绝对位姿。The positioning module is configured to determine, according to the target image, the absolute pose of the camera that collects the image to be processed in the world coordinate system.
- 根据权利要求10所述的装置,其特征在于,所述获取模块包括:The device according to claim 10, wherein the acquisition module comprises:分割子模块,用于对预定图像集合中的每张图像进行像素级语义分割;The segmentation sub-module is used to perform pixel-level semantic segmentation on each image in the predetermined image set;第一确定子模块,用于根据像素级语义分割的结果确定所述每张图像中属于可移动物体的第一像素点和属于背景的第二像素点;The first determining submodule is configured to determine the first pixel that belongs to the movable object and the second pixel that belongs to the background in each image according to the result of pixel-level semantic segmentation;第二确定子模块,用于基于所述每张图像中所述第一像素点和所述第二像素点的统计分布,确定与所述预定图像集合中的图像等大的图像模板包括的多个像素点中每个像素点处出现所述可移动物体的所述先验概率。The second determining sub-module is configured to determine, based on the statistical distribution of the first pixel points and the second pixel points in each image, how many image templates that are as large as the images in the predetermined image set include The prior probability that the movable object appears at each pixel in each pixel.
- 根据权利要求10或11所述的装置,其特征在于,所述执行模块包括:The device according to claim 10 or 11, wherein the execution module comprises:采样子模块,用于对所述待处理图像所包括的至少部分像素点对应的先验概率进行采样;The sampling sub-module is used to sample the prior probabilities corresponding to at least some pixels included in the image to be processed;执行子模块,用于在所述待处理图像上去除先验概率的采样值大于预设阈值的像素点,得到所述目标图像。The execution sub-module is used to remove the pixel points whose a priori probability sampling value is greater than a preset threshold on the image to be processed to obtain the target image.
- 根据权利要求12所述的装置,其特征在于,采样次数为多次时,执行丢弃部分像素点的操作之后得到的多个目标图像两两之间存在至少一个不同的像素点。The device according to claim 12, wherein when the number of sampling times is multiple times, there is at least one different pixel point between the multiple target images obtained after the operation of discarding some pixels.
- 根据权利要求10-13任一项所述的装置,其特征在于,所述定位模块包括:The device according to any one of claims 10-13, wherein the positioning module comprises:第一处理子模块,用于经神经网络提取所述目标图像中的特征参数,得到特征提取图像;The first processing sub-module is used to extract feature parameters in the target image via a neural network to obtain a feature extraction image;第二处理子模块,用于在所述神经网络的预设空间维度和/或预设通道维度上,增加所述特征提取图像中属于背景的第二像素点所对应的权重值;The second processing sub-module is configured to increase the weight value corresponding to the second pixel point belonging to the background in the feature extraction image in the preset space dimension and/or the preset channel dimension of the neural network;第一定位子模块,用于经所述神经网络对权重值调整后的特征提取图像进行分析,得到采集所述待处理图像的相机在世界坐标系下的所述绝对位姿。The first positioning sub-module is configured to analyze the feature extraction image after the weight value is adjusted by the neural network to obtain the absolute pose of the camera that collects the image to be processed in the world coordinate system.
- 根据权利要求14所述的装置,其特征在于,所述装置还包括:The device according to claim 14, wherein the device further comprises:训练模块,用于根据所述绝对位姿和预先确定的采集所述待处理图像的所述相机的位姿真值的差异,调整所述神经网络的网络参数,训练得到目标神经网络。The training module is configured to adjust the network parameters of the neural network according to the difference between the absolute pose and the predetermined true value of the camera that collects the image to be processed, to train the target neural network.
- 根据权利要求15所述的装置,其特征在于,所述定位模块包括:The device according to claim 15, wherein the positioning module comprises:第二定位子模块,用于将所述待处理图像输入所述目标神经网络,得到所述待处理图像的相机在世界坐标系下的所述绝对位姿。The second positioning sub-module is used to input the image to be processed into the target neural network to obtain the absolute pose of the camera of the image to be processed in the world coordinate system.
- 根据权利要求10-16任一项所述的装置,其特征在于,所述待处理图像包括所述相机采集的具有时序性的至少两帧图像;The device according to any one of claims 10-16, wherein the image to be processed comprises at least two frames of images with time series collected by the camera;所述装置还包括:The device also includes:第一确定模块,用于根据所述至少两帧图像确定所述相机在拍摄所述至少两帧图像时的相对位姿;The first determining module is configured to determine the relative pose of the camera when shooting the at least two frames of images according to the at least two frames of images;第二确定模块,用于根据所述相机的相对位姿和所述绝对位姿,确定所述相机的修正位姿。The second determining module is configured to determine the corrected pose of the camera according to the relative pose of the camera and the absolute pose.
- 根据权利要求17所述的装置,其特征在于,所述第二确定模块还包括:The device according to claim 17, wherein the second determining module further comprises:第三确定子模块,用于确定所述绝对位姿的确定性概率;The third determining sub-module is used to determine the certainty probability of the absolute pose;第四确定子模块,用于根据所述确定性概率确定所述相对位姿的第一权重和所述绝对位姿的第二权重;A fourth determination submodule, configured to determine the first weight of the relative pose and the second weight of the absolute pose according to the certainty probability;第五确定子模块,用于根据所述相对位姿、所述第一权重、所述绝对位姿和所述第二权重,确定所述相机的修正位姿。The fifth determining sub-module is configured to determine the corrected pose of the camera according to the relative pose, the first weight, the absolute pose, and the second weight.
- 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,所述计算机程序用于执行上述权利要求1-9任一所述的相机定位方法。A computer-readable storage medium, the computer-readable storage medium stores a computer program, and the computer program is used to execute the camera positioning method according to any one of claims 1-9.
- 一种电子设备,包括:An electronic device including:处理器;processor;用于存储所述处理器可执行指令的存储器;A memory for storing executable instructions of the processor;其中,所述处理器用于调用所述存储器中存储的可执行指令,实现权利要求1-9中任一项所述的相机定位方法。Wherein, the processor is configured to call executable instructions stored in the memory to implement the camera positioning method according to any one of claims 1-9.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2021534170A JP2022513868A (en) | 2019-05-27 | 2020-05-22 | Camera positioning |
KR1020217019918A KR20210095925A (en) | 2019-05-27 | 2020-05-22 | camera positioning |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910447759.7A CN112001968B (en) | 2019-05-27 | 2019-05-27 | Camera positioning method and device and storage medium |
CN201910447759.7 | 2019-05-27 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020238790A1 true WO2020238790A1 (en) | 2020-12-03 |
Family
ID=73461260
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/091768 WO2020238790A1 (en) | 2019-05-27 | 2020-05-22 | Camera positioning |
Country Status (4)
Country | Link |
---|---|
JP (1) | JP2022513868A (en) |
KR (1) | KR20210095925A (en) |
CN (1) | CN112001968B (en) |
WO (1) | WO2020238790A1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112885134A (en) * | 2021-01-24 | 2021-06-01 | 成都智慧赋能科技有限公司 | Big data-based smart city traffic management method |
CN114118367A (en) * | 2021-11-16 | 2022-03-01 | 上海脉衍人工智能科技有限公司 | Method and equipment for constructing incremental nerve radiation field |
CN114693776A (en) * | 2022-03-25 | 2022-07-01 | 广东电网有限责任公司 | Cable position information determining method, device, equipment and storage medium |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104978722A (en) * | 2015-07-06 | 2015-10-14 | 天津大学 | Multi-exposure image fusion ghosting removing method based on background modeling |
CN105931275A (en) * | 2016-05-23 | 2016-09-07 | 北京暴风魔镜科技有限公司 | Monocular and IMU fused stable motion tracking method and device based on mobile terminal |
CN108257177A (en) * | 2018-01-15 | 2018-07-06 | 天津锋时互动科技有限公司深圳分公司 | Alignment system and method based on space identification |
CN109387204A (en) * | 2018-09-26 | 2019-02-26 | 东北大学 | The synchronous positioning of the mobile robot of dynamic environment and patterning process in faced chamber |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5333860B2 (en) * | 2010-03-31 | 2013-11-06 | アイシン・エィ・ダブリュ株式会社 | Vehicle position detection system using landscape image recognition |
JP5743849B2 (en) * | 2011-10-27 | 2015-07-01 | 株式会社日立製作所 | Video analysis apparatus and system |
JP2016177388A (en) * | 2015-03-18 | 2016-10-06 | 株式会社リコー | Mobile object position/attitude measuring apparatus |
JP6985897B2 (en) * | 2017-01-06 | 2021-12-22 | キヤノン株式会社 | Information processing equipment and its control method, program |
US11348274B2 (en) * | 2017-01-23 | 2022-05-31 | Oxford University Innovation Limited | Determining the location of a mobile device |
US10467756B2 (en) * | 2017-05-14 | 2019-11-05 | International Business Machines Corporation | Systems and methods for determining a camera pose of an image |
JP7043755B2 (en) * | 2017-08-29 | 2022-03-30 | ソニーグループ株式会社 | Information processing equipment, information processing methods, programs, and mobiles |
CN107833236B (en) * | 2017-10-31 | 2020-06-26 | 中国科学院电子学研究所 | Visual positioning system and method combining semantics under dynamic environment |
-
2019
- 2019-05-27 CN CN201910447759.7A patent/CN112001968B/en active Active
-
2020
- 2020-05-22 JP JP2021534170A patent/JP2022513868A/en active Pending
- 2020-05-22 WO PCT/CN2020/091768 patent/WO2020238790A1/en active Application Filing
- 2020-05-22 KR KR1020217019918A patent/KR20210095925A/en not_active Application Discontinuation
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104978722A (en) * | 2015-07-06 | 2015-10-14 | 天津大学 | Multi-exposure image fusion ghosting removing method based on background modeling |
CN105931275A (en) * | 2016-05-23 | 2016-09-07 | 北京暴风魔镜科技有限公司 | Monocular and IMU fused stable motion tracking method and device based on mobile terminal |
CN108257177A (en) * | 2018-01-15 | 2018-07-06 | 天津锋时互动科技有限公司深圳分公司 | Alignment system and method based on space identification |
CN109387204A (en) * | 2018-09-26 | 2019-02-26 | 东北大学 | The synchronous positioning of the mobile robot of dynamic environment and patterning process in faced chamber |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112885134A (en) * | 2021-01-24 | 2021-06-01 | 成都智慧赋能科技有限公司 | Big data-based smart city traffic management method |
CN112885134B (en) * | 2021-01-24 | 2023-05-16 | 陕西合友网络科技有限公司 | Smart city traffic management method based on big data |
CN114118367A (en) * | 2021-11-16 | 2022-03-01 | 上海脉衍人工智能科技有限公司 | Method and equipment for constructing incremental nerve radiation field |
CN114118367B (en) * | 2021-11-16 | 2024-03-29 | 上海脉衍人工智能科技有限公司 | Method and equipment for constructing incremental nerve radiation field |
CN114693776A (en) * | 2022-03-25 | 2022-07-01 | 广东电网有限责任公司 | Cable position information determining method, device, equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
KR20210095925A (en) | 2021-08-03 |
JP2022513868A (en) | 2022-02-09 |
CN112001968B (en) | 2022-07-15 |
CN112001968A (en) | 2020-11-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109784333B (en) | Three-dimensional target detection method and system based on point cloud weighted channel characteristics | |
WO2020238790A1 (en) | Camera positioning | |
US20200051250A1 (en) | Target tracking method and device oriented to airborne-based monitoring scenarios | |
WO2021196294A1 (en) | Cross-video person location tracking method and system, and device | |
CN110570454B (en) | Method and device for detecting foreign matter invasion | |
EP2798611B1 (en) | Camera calibration using feature identification | |
CN105718872B (en) | Auxiliary method and system for rapidly positioning lanes on two sides and detecting vehicle deflection angle | |
US11748894B2 (en) | Video stabilization method and apparatus and non-transitory computer-readable medium | |
CN111126399A (en) | Image detection method, device and equipment and readable storage medium | |
CN105930822A (en) | Human face snapshot method and system | |
CN103824070A (en) | Rapid pedestrian detection method based on computer vision | |
CN111126184B (en) | Post-earthquake building damage detection method based on unmanned aerial vehicle video | |
Ringwald et al. | UAV-Net: A fast aerial vehicle detector for mobile platforms | |
CN110569754A (en) | Image target detection method, device, storage medium and equipment | |
WO2021218671A1 (en) | Target tracking method and device, and storage medium and computer program | |
US20190114753A1 (en) | Video Background Removal Method | |
CN110443279B (en) | Unmanned aerial vehicle image vehicle detection method based on lightweight neural network | |
CN113191180A (en) | Target tracking method and device, electronic equipment and storage medium | |
CN104966095A (en) | Image target detection method and apparatus | |
CN111738033A (en) | Vehicle driving information determination method and device based on plane segmentation and vehicle-mounted terminal | |
Mejia et al. | Vehicle speed estimation using computer vision and evolutionary camera calibration | |
Miller et al. | Person tracking in UAV video | |
CN104715476A (en) | Salient object detection method based on histogram power function fitting | |
CN114071015A (en) | Method, device, medium and equipment for determining linkage snapshot path | |
CN113591735A (en) | Pedestrian detection method and system based on deep learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20812552 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2021534170 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20217019918 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20812552 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20812552 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 27.06.2022) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20812552 Country of ref document: EP Kind code of ref document: A1 |