CN117495698A - Flying object identification method, system, intelligent terminal and computer readable storage medium - Google Patents

Flying object identification method, system, intelligent terminal and computer readable storage medium Download PDF

Info

Publication number
CN117495698A
CN117495698A CN202410001150.8A CN202410001150A CN117495698A CN 117495698 A CN117495698 A CN 117495698A CN 202410001150 A CN202410001150 A CN 202410001150A CN 117495698 A CN117495698 A CN 117495698A
Authority
CN
China
Prior art keywords
flying object
image frame
image
identified
coordinate system
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202410001150.8A
Other languages
Chinese (zh)
Inventor
李家悦
郑明炀
苏小杭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fujian Zhuohang Special Equipment Co ltd
Original Assignee
Fujian Zhuohang Special Equipment Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fujian Zhuohang Special Equipment Co ltd filed Critical Fujian Zhuohang Special Equipment Co ltd
Priority to CN202410001150.8A priority Critical patent/CN117495698A/en
Publication of CN117495698A publication Critical patent/CN117495698A/en
Pending legal-status Critical Current

Links

Abstract

The application relates to the technical field of computer vision, and provides a method, a system, an intelligent terminal and a computer readable storage medium for identifying flying objects, wherein the method comprises the following steps: splicing the first image frame and the second image frame to obtain a scene fusion image; inputting the scene fusion image into an identification model to obtain an identification flying object output by the identification model; and acquiring the space coordinates of the identified flying object under a phase coordinate system based on the two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame respectively. The application aims at the intelligent terminal with smaller carrier volume and smaller power consumption, can rapidly identify the flying object under the condition of no manual intervention, and expands the stereoscopic vision reconstruction work about identifying the flying object, so that the recognition precision and efficiency are considered under the condition of low cost, the working intensity of monitoring and the like is further reduced, and the working efficiency is improved.

Description

Flying object identification method, system, intelligent terminal and computer readable storage medium
Technical Field
The present disclosure relates to the field of computer vision, and in particular, to a method and system for identifying a flying object, an intelligent terminal, and a computer readable storage medium.
Background
The identification and indication of the flying object is an important function of security monitoring, mixed (augmented) reality and other applications. The intelligent and rapid recognition of the flying object and the positioning indication of the flying object are the basis for the next treatment work. The aircraft that typically need to be identified include personnel, vehicles, buildings, and the like. At present, a video or an image shot by a camera is mainly adopted for identifying the flying object, and then the flying object is calibrated by matching with laser and radar ranging. Such schemes are limited by two factors, namely, when using two cameras or multiple cameras, because the different cameras recognize the fine difference of the position of the flying object, the flying object may be repeatedly positioned. Secondly, the radiation distance of a ranging system such as laser and radar and the power consumption of the ranging system are positively correlated. Therefore, in consideration of the size and the power consumption of the carrier, the implementation of the flyer identification scheme in the intelligent wearable device or the intelligent mobile phone is generally realized by adopting the combination of an RGB camera and a depth camera or a short-distance laser radar, and the identification precision and the identification efficiency are lower due to the smaller perception range.
Disclosure of Invention
The application provides a flying object identification method, a system, an intelligent terminal and a computer readable storage medium, which are used for solving the defect that in the prior art, aiming at the setting with smaller power consumption, the identification precision and the efficiency are lower due to smaller perception range.
In a first aspect, the present application provides a method for identifying a flying object, which is applied to an intelligent terminal, and includes:
splicing the first image frame and the second image frame to obtain a scene fusion image; the first image frame and the second image frame are respectively obtained by shooting by two cameras with parallel optical axes in the binocular camera;
inputting the scene fusion image into an identification model to obtain an identification flying object output by the identification model; the identification model is obtained by training the identification flying object type label marked by the sample scene image based on the sample scene image;
acquiring space coordinates of the identified flying object under a phase coordinate system based on two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame respectively; the phase plane coordinate system is established according to a base line formed between two cameras with parallel optical axes in the binocular camera.
According to the method for identifying the flying object provided by the application, the method for acquiring the space coordinates of the identified flying object under the phase plane coordinate system based on the two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame respectively comprises the following steps:
Determining parallax information based on the X-axis coordinate value in the two-dimensional image coordinates of the identified flying object and the depth information of the identified flying object;
and determining the space coordinates of the identified flying object under a phase plane coordinate system based on the two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame respectively and the parallax information.
According to the method for identifying the flying object provided by the application, the step of splicing the first image frame and the second image frame to obtain the scene fusion image comprises the following steps:
respectively extracting features of the first image frame and the second image frame to obtain a first feature point set and a second feature point set;
and matching based on the first characteristic point set and the second characteristic point set to obtain the scene fusion image.
According to the method for identifying the flying object provided by the application, after the space coordinates of the identified flying object under the phase plane coordinate system are obtained, the method further comprises the following steps:
generating a three-dimensional space model of the identified flying object under a camera coordinate system based on the first feature point set and the second feature point set, and the first feature depth information and the second feature depth information;
Acquiring the position coordinates of the identified flying object relative to a camera according to the space coordinates of the identified flying object under the phase plane coordinate system and the three-dimensional space model of the identified flying object under the camera coordinate system;
based on the position coordinates of the identified flying object relative to the camera and the camera gesture data, calculating the position coordinates of the identified flying object under an earth coordinate system;
the camera coordinate system is established by taking the intelligent terminal as a coordinate origin; the first characteristic depth information comprises depth information corresponding to each characteristic point in the first characteristic point set; the second feature depth information includes depth information corresponding to each feature point in the second feature point set.
According to the method for identifying the flying object provided by the application, the generating of the three-dimensional space model of the identified flying object under the camera coordinate system based on the first feature point set, the second feature point set, the first feature depth information and the second feature depth information comprises the following steps:
generating sparse point cloud information based on the first feature point set and the second feature point set, and the first feature depth information and the second feature depth information;
And encrypting and smoothing the sparse point cloud information to form the three-dimensional space model of the identified flying object under the camera coordinate system.
According to the method for identifying the flying object, the identification model comprises a feature extraction layer and a classification layer;
correspondingly, the step of inputting the scene fusion image into the recognition model to acquire the recognized flying object output by the recognition model comprises the following steps:
inputting the scene fusion image into the feature extraction layer to obtain a feature vector output by the feature extraction layer;
and inputting the feature vector to the classification layer, and processing by adopting a multi-head attention mechanism to obtain the identified flying object output by the classification layer.
According to the method for identifying the flying object provided by the application, before the first image frame and the second image frame are spliced to obtain the scene fusion image, the method further comprises the following steps:
and obtaining an internal reference matrix of the binocular camera through calibration so as to calculate the position coordinate of the identified flying object under the earth coordinate system based on the position coordinate of the identified flying object relative to the camera, the camera gesture data and the internal reference matrix of the binocular camera.
In a second aspect, the present application further provides a flyer identification system, provided in an intelligent terminal, the system including:
the binocular fusion module is used for splicing the first image frame and the second image frame to obtain a scene fusion image; the first image frame and the second image frame are respectively obtained by shooting by two cameras with parallel optical axes in the binocular camera;
the flyer identification module is used for inputting the scene fusion image into an identification model to acquire an identified flyer output by the identification model; the recognition model is obtained by training a recognition flyer category label marked by a sample scene image;
the flying object positioning module is used for acquiring space coordinates of the identified flying object under a phase plane coordinate system based on two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame respectively; the phase plane coordinate system is established according to a base line formed between two cameras with parallel optical axes in the binocular camera.
In a third aspect, the present application further provides an intelligent terminal, including an intelligent terminal body provided with a binocular camera, and a vision processor communicatively connected with the binocular camera; the visual processor executes a program to implement the method for identifying the flying object according to any one of the above;
The intelligent terminal body comprises a mobile phone terminal, augmented reality glasses, mixed reality glasses and augmented reality glasses.
In a fourth aspect, the present application also provides a non-transitory computer readable storage medium having stored thereon a computer program which, when executed by a processor, implements a method of identifying a flying object as described in any one of the above.
According to the method and the system for identifying the flying object, after the first image frame and the second image frame which are respectively acquired by the binocular camera are fused, only the scene fusion image which is reserved at the overlapped part is used as the input of the identification model, and the output result is the identification flying object which is correspondingly framed in the image, so that the two-dimensional image coordinates of the identification flying object in the first image frame and the second image frame are utilized to convert the space coordinates of the identification flying object in the phase plane coordinate system to carry out stereoscopic vision reconstruction. Aiming at the intelligent terminal with smaller carrier volume and power consumption, the flying object can be rapidly identified under the condition of no manual intervention, the stereoscopic vision reconstruction work about the identification of the flying object is developed, the identification precision and the efficiency are considered under the condition of low cost, the working intensity of monitoring and the like is further reduced, and the working efficiency is improved.
Drawings
For a clearer description of the present application or of the prior art, the drawings that are used in the description of the embodiments or of the prior art will be briefly described, it being apparent that the drawings in the description below are some embodiments of the present application, and that other drawings may be obtained from these drawings without inventive effort for a person skilled in the art.
Fig. 1 is a schematic flow chart of a method for identifying a flying object provided by the application.
Fig. 2 is a schematic structural diagram of the flyer identification system provided in the present application.
Fig. 3 is a schematic structural diagram of an intelligent terminal provided in the present application.
Detailed Description
For the purposes of making the objects, technical solutions and advantages of the present application more apparent, the technical solutions in the present application will be clearly and completely described below with reference to the drawings in the present application, and it is apparent that the described embodiments are some, but not all, embodiments of the present application. All other embodiments, which can be made by one of ordinary skill in the art based on the embodiments herein without making any inventive effort, are intended to be within the scope of the present application.
The terms "first," "second," and the like in the description of the present application, are used for distinguishing between similar objects and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used may be interchanged, as appropriate, such that embodiments of the present application may be implemented in sequences other than those illustrated or described herein, and that the objects identified by "first," "second," etc. are generally of a type and not limited to the number of objects, e.g., the first object may be one or more. It is to be understood that the terminology used in the description of the present application is for the purpose of describing particular embodiments only and is not intended to be limiting of the application. As used in this specification, the singular forms "a", "an" and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. The terms "comprises" and "comprising" indicate the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
Optionally, fig. 1 is a schematic flow chart of a method for identifying a flying object provided in the present application. As shown in fig. 1, the method for identifying a flying object provided in the embodiment of the present application is applied to an intelligent terminal, and the method includes steps 101 to 103:
And step 101, splicing the first image frame and the second image frame to obtain a scene fusion image. The first image frame and the second image frame are respectively shot by two cameras with parallel optical axes in the binocular camera.
It should be noted that, the execution subject of the flyer identification method is a flyer identification system, and the system may be integrated on a control chip in the intelligent terminal.
The flyer identification method is suitable for identifying the flyers through the electronic equipment with the flyer identification system, and further identifying the true geographic positions of the flyer ground objects.
The electronic device described above may be implemented in various forms. For example, the electronic device described in the embodiments of the present application may be a smart terminal device integrating a flyer recognition system and a video capture system, such as a mobile terminal of a mobile phone, a smart phone, a notebook computer, a digital broadcast receiver, a PDA (personal digital assistant), a PAD (tablet computer), a PMP (portable multimedia player), a navigation system, a smart bracelet, a smart watch, a digital camera, and the like.
The electronic device described in the embodiments of the present application may also be a terminal device provided with a flyer identification system, where the terminal needs to be connected with the video acquisition system in a communication manner. A fixed terminal such as a desktop computer or the like. In the following, it is assumed that the electronic device is a mobile terminal. However, it will be understood by those skilled in the art that the configuration according to the embodiment of the present application can be applied to a fixed type terminal in addition to elements particularly used for a moving purpose.
The video acquisition system can be combined according to the installation and use characteristics of the intelligent terminal, for example, the two RGB cameras are parallel in optical axis and are rigidly connected, and the two RGB cameras are integrated into a binocular camera with fixed relative position between the two cameras.
The first image frame refers to a video image acquired in real time by an RGB camera on the left side of the binocular camera.
The second image frame refers to a video image acquired in real time by the RGB camera on the right side in the binocular camera.
Specifically, the flyer recognition system performs feature extraction on a first image frame and a second image frame which are acquired by the binocular camera at the same time, and performs registration by using the extracted features, so that a scene fusion image with non-overlapping portions in the two images removed can be obtained.
102, inputting the scene fusion image into a recognition model, and obtaining a recognized flying object output by the recognition model. The recognition model is obtained by training a recognition flying object type label marked on the sample scene image based on the sample scene image.
It should be noted that the recognition model may be a neural network model, and the structure and parameters of the neural network include, but are not limited to, the input layer, the hidden layer, the number of output layers, and the weight parameters of each layer. The kind and structure of the neural network are not particularly limited in the embodiments of the present application. For example, the recognition model may be a neural network model consisting of an input layer, a hidden layer, and an output layer, wherein:
the input layer directly receives and digitally encodes the scene fusion image at the forefront portion of the overall network.
The hidden layer may have one or more layers and the input vector is calculated by means of weighted summation of its neurons.
The output layer is the last layer and is used for decoding the vector obtained after the weighted summation, and different classes of flying objects are framed from the scene fusion image.
The sample data includes a sample scene image corresponding to the sample data, and an identification flying object type label marked on the sample scene image. The sample data is divided into a training set and a testing set according to a certain proportion.
Illustratively, the duty cycle of the training set and the test set in the sample data includes, but is not limited to, 9: 1. 8:2, etc., to which the embodiments of the present application are not particularly limited.
In step 102, the flyer identification system initializes the weight coefficient between each layer of the constructed identification model, inputs a set of sample scene images in the training set and the identification flyer category labels marked on the sample scene images to the neural network under the current weight coefficient, and sequentially calculates the output of each node of the input layer, the hidden layer and the output layer. And correcting the weight coefficient between each node of the input layer and the hidden layer according to the accumulated error between the final output result of the output layer and the actual connection position state type of the output layer and the gradient descent method. According to the above process, until all samples in the training set are traversed, the weight coefficients of the input layer and the hidden layer can be obtained.
And the flyer identification system restores the identification model in the step 102 according to the weight coefficients of the neural network input layer and the hidden layer, and inputs the scene fusion image obtained in the step 101 into the trained identification model, so that the identified flyers of various categories identified by the scene fusion image can be obtained.
Step 103, based on the two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame, acquiring the space coordinates of the identified flying object under a phase plane coordinate system.
The phase plane coordinate system is established according to a base line formed between two cameras with parallel optical axes in the binocular camera.
Specifically, in step 103, after the flying object identification system specifically identifies the flying objects, the flying object identification system respectively obtains two-dimensional image coordinates of the center of each identified flying object in the image coordinate system where the first image frame and the second image frame are respectively located, and based on the two-dimensional image coordinates, converts the spatial coordinates of the identified flying object in the phase plane coordinate system, so as to perform stereoscopic imaging on the binocular camera.
The phase plane coordinate system is a virtual imaging plane created for a binocular camera, the base line formed between two cameras with parallel optical axes in the binocular camera extends to form an X axis, the left boundary of a picture shot by a left camera extends to form a Y axis, and finally a straight line perpendicular to an XY plane passing through an intersection point between the X axis and the Y axis is taken as a Z axis.
After the first image frame and the second image frame which are respectively acquired by the binocular camera are fused, a scene fusion image which only keeps an overlapped part is used as input of an identification model, and an output result is an identification flying object which corresponds to a frame in the image, so that the two-dimensional image coordinates of the identification flying object in the first image frame and the second image frame are utilized to convert the space coordinates of the identification flying object in a phase plane coordinate system to carry out stereoscopic vision reconstruction. Aiming at the intelligent terminal with smaller carrier volume and power consumption, the flying object can be rapidly identified under the condition of no manual intervention, the stereoscopic vision reconstruction work about the identification of the flying object is developed, the identification precision and the efficiency are considered under the condition of low cost, the working intensity of monitoring and the like is further reduced, and the working efficiency is improved.
On the basis of any one of the above embodiments, acquiring spatial coordinates of the identified flying object in a phase plane coordinate system based on two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame, respectively, includes: and determining parallax information based on the X-axis coordinate value in the two-dimensional image coordinates of the identified flying object and the depth information of the identified flying object.
Specifically, in step 103, assuming that the C1 camera coordinate system in which the first image frame is located is O1x1y1z1, the C2 camera coordinate system in which the second image frame is located is O2x2y2z2, the focal length is f, the distance between the two cameras of the binocular camera is d, the center point P of the identified flying object is (x 1, y1, z 1) in the C1 camera coordinate system, and is (x 2, y2, z 2) in the C2 camera coordinate system. The coordinates of the optical centers of the left camera of the binocular camera in the phase plane coordinate system are (u 1, v1, 0), and the coordinates of the optical centers of the right camera of the binocular camera in the phase plane coordinate system are (u 2, v2, 0).
According to the principle of triangle, the following deduction relation exists for the left camera and the right camera:
assuming that, ideally, there is parallax between the C1 camera coordinate system and the C2 camera coordinate system only in the X-axis where the base line is located, the relationship of the phase plane coordinate system ozz to the C1 camera coordinate system and the C2 camera coordinate system can be expressed as follows:
Two formulas can be obtained on the same time:
the available parallax information may then be calculated by the following formula:
wherein,to identify depth information of the flyer.
And determining the space coordinates of the identified flying object under a phase plane coordinate system based on the two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame respectively and the parallax information.
Specifically, the flyer identification system combines the formula (4) and the formula (1), and converts the space coordinates (X, Y, Z) of the identified flyer under the phase plane coordinate system, wherein the conversion formula is as follows:
according to the method and the device for identifying the three-dimensional coordinates of the flying object, the parallax information is determined through identifying the two-dimensional image coordinates of the flying object in the first image frame and the second image frame respectively, and the three-dimensional coordinates of the flying object can be recovered by utilizing the parallax information and the image point coordinates of the flying object which are matched in the first image frame and the second image frame respectively. The visual field range of the camera of the intelligent terminal can be expanded, so that the flexibility of the vision system is enhanced.
On the basis of any one of the above embodiments, stitching the first image frame and the second image frame to obtain a scene fusion image includes: and respectively carrying out feature extraction on the first image frame and the second image frame to obtain a first feature point set and a second feature point set.
Specifically, in step 101, the flyer recognition system performs feature point extraction on the first image frame, and integrates feature points that do not change with rotation, scale, and illumination changes into a first feature point set. And simultaneously, extracting the characteristic points of the second image frame to obtain a second characteristic point set.
Among them, the feature point extraction is an algorithm used, including but not limited to a scale-invariant feature transform (Scale Invariant Feature Transform, SIFT) algorithm, a Speeded-up robust feature (Speeded-Up Robust Features, SURF) algorithm, and a fast feature point extraction and description (Oriented FAST and Rotated BRIEF, ORB), which is not specifically limited in this embodiment of the present application.
And matching based on the first characteristic point set and the second characteristic point set to obtain the scene fusion image.
Specifically, the flyer identification system finds out matched feature point pairs by carrying out similarity measurement on the first feature point set and the second feature point set, and after the spatial mapping relation between two image frames is determined, the overlapped parts of two images with the same plane in space can be associated through homography transformation to obtain a scene fusion image.
It should be noted that, for the application with higher response time requirement, image fusion may not be performed, only one of the binocular cameras is selected as a basis, and the binocular overlapping portion is reserved and then used as an input image of the recognition model.
According to the method and the device for identifying the scene fusion image, the respective characteristic points are extracted from the first image frame and the second image frame, the characteristic point sets of the two images are matched, the spatial mapping relation between the two images is optimized by affine transformation/perspective transformation and the like after optimal matching is obtained, one of the images is deformed into the same spatial layout as the other image, and then the images are spliced into the scene fusion image, so that identification accuracy is guaranteed.
On the basis of any one of the foregoing embodiments, after the acquiring the spatial coordinates of the identified flying object in the phase plane coordinate system, the method further includes: and generating a three-dimensional space model of the identified flying object under a camera coordinate system based on the first feature point set and the second feature point set, and the first feature depth information and the second feature depth information.
The camera coordinate system is established by taking the intelligent terminal as a coordinate origin; the first characteristic depth information comprises depth information corresponding to each characteristic point in the first characteristic point set; the second feature depth information includes depth information corresponding to each feature point in the second feature point set.
It should be noted that, the flyer identification system also needs to establish communication connection with the ranging sensing system to obtain the first feature depth information, where the first feature depth information includes depth information corresponding to each feature point in the first feature point set, and integrate the first feature depth information. And similarly, obtaining second characteristic depth information corresponding to the second characteristic point set.
Specifically, after step 103, the flyer identification system finds a matched pair of feature points by performing similarity measurement on the first feature point set and the second feature point set, and then respectively combines depth values recorded in the first feature depth information and depth values recorded in the second feature depth information with the matched pair of feature points to construct a three-dimensional space model for identifying the flyer under a camera coordinate system established by taking the intelligent terminal as an origin.
And acquiring the position coordinates of the identified flying object relative to the camera according to the space coordinates of the identified flying object under the phase plane coordinate system and the three-dimensional space model of the identified flying object under the camera coordinate system.
Specifically, the flying object recognition system obtains the position coordinates of the recognition flying object relative to the camera by combining the spatial coordinates of the recognition flying object in the phase plane coordinate system obtained in the step 103 and converting the spatial coordinates recorded in the three-dimensional space model.
And calculating the position coordinate of the identified flying object under the earth coordinate system based on the position coordinate of the identified flying object relative to the camera and the camera gesture data.
Specifically, the flying object identification system is matched with Beidou, GPS, magnetic compass and inertial measurement device to carry out coordinate system conversion on the position coordinates of the identification flying object relative to the camera according to the camera gesture data acquired by the binocular camera, so that the position coordinates of the identification flying object under the earth coordinate system can be obtained through calculation.
According to the method and the device, based on the matched characteristic points between binocular images and corresponding depth information, after a three-dimensional space model for identifying the flying object under a camera coordinate system is formed, the position coordinate of the flying object relative to a camera is converted through the space coordinate of the flying object under a phase plane coordinate system and the coordinate of the three-dimensional space model for identifying the flying object under the camera coordinate system, and then the position coordinate of the flying object under the earth coordinate system is calculated by combining the camera gesture data. The intelligent terminal can be used for positioning and real-time tracking of the flying object, and can also be used for identifying the flying object aiming at a visual blind area, so that the situation awareness capability of a terminal user is improved.
On the basis of any one of the above embodiments, generating the three-dimensional space model of the identified flying object under the camera coordinate system based on the first feature point set and the second feature point set, and the first feature depth information and the second feature depth information includes: generating sparse point cloud information based on the first feature point set and the second feature point set, and the first feature depth information and the second feature depth information.
Specifically, the flyer identification system finds matched feature point pairs in the first feature point set and the second feature point set by performing similarity measurement, and depth values recorded in the first feature depth information and the second feature depth information respectively serve as projections of real three-dimensional space points on an image plane. And performing triangulation (triangulation) on the feature points, constructing a triangle by using geometric information (epipolar geometry) to determine the position of the three-dimensional space point, and performing second projection on the pose of the camera, namely re-projecting the calculated virtual pixel point coordinates. And taking the coordinate value difference of the two as a reprojection error to minimize the sum of reprojection errors of the matching point pairs to obtain the optimal camera pose parameter and the coordinates of the three-dimensional space points, thereby forming sparse point cloud information.
And encrypting and smoothing the sparse point cloud information to form the three-dimensional space model of the identified flying object under the camera coordinate system.
Specifically, the flying object recognition system encrypts and smoothes the sparse point cloud information, eliminates discontinuous surfaces such as holes and the like, and forms a three-dimensional space model for recognizing the flying object under a camera coordinate system.
According to the embodiment of the application, sparse point cloud information is generated based on the matched characteristic points between the binocular images and the corresponding depth information, and encryption and smoothing processing are carried out on the sparse point cloud information, so that a three-dimensional space model for reserving and identifying foreground points of flying objects is formed. The cost function is constructed to optimize through the error obtained by comparing the pixel coordinates (with the position obtained by projecting the 3D point according to the current estimated pose), so that the calculation error of the homography matrix is considered, the measurement error of the image point is also considered, and the positioning precision of the flying object is improved.
On the basis of any one of the above embodiments, the recognition model includes a feature extraction layer and a classification layer.
Specifically, the hidden layer of the recognition model includes a feature extraction layer and a classification layer, wherein:
and the feature extraction layer is used for reducing the dimension of the scene fusion image.
And the classification layer is used for mapping the feature vectors after the dimension reduction to different aircraft categories.
Correspondingly, the step of inputting the scene fusion image into the recognition model to acquire the recognized flying object output by the recognition model comprises the following steps: and inputting the scene fusion image to the feature extraction layer to obtain the feature vector output by the feature extraction layer.
Specifically, in step 102, the flyer recognition system extracts features from the scene fusion image by using the convolutional neural network deployed in the feature extraction layer, and obtains feature vectors after dimension reduction.
And inputting the feature vector to the classification layer, and processing by adopting a multi-head attention mechanism to obtain the identified flying object output by the classification layer.
Specifically, the flyer recognition system inputs the feature vector output by the feature extraction layer to a transducer network deployed by the classification layer, and further learns global features by utilizing a multi-head attention mechanism of the transducer network, so that each point recorded in the feature vector can interact with other points to recognize flyers of one category in an image, and recognition of other flyers can be reduced to a category with semantic association according to semantic description of the flyers until all flyers are recognized.
According to the embodiment of the application, the neural network model is utilized to extract the characteristics of the scene fusion image, and the obtained characteristic vectors are subjected to classification mapping by adopting a multi-head attention mechanism, so that different types of identified flying objects are obtained. Global information can be effectively obtained through a multi-head attention mechanism and can be mapped to a plurality of spaces, so that the expression capacity of the model is enhanced, and the efficiency of identifying flying objects is improved.
On the basis of any one of the above embodiments, before the splicing the first image frame and the second image frame to obtain the scene fusion image, the method further includes: and obtaining an internal reference matrix of the binocular camera through calibration so as to calculate the position coordinate of the identified flying object under the earth coordinate system based on the position coordinate of the identified flying object relative to the camera, the camera gesture data and the internal reference matrix of the binocular camera.
Specifically, prior to step 101, the flyer identification uses Zhang Youzheng calibration method or other calibration method to determine an internal reference matrix of each camera in the binocular camera, so as to be used as an initial estimated value of the camera pose in subsequent orthodontic and coordinate system conversion.
According to the embodiment of the application, calibration correction is advanced before the recognition of the flying object, barrel-shaped and pillow-shaped distortion of the shot image can be avoided, and the accuracy of the recognition of the flying object is improved.
Fig. 2 is a schematic structural diagram of the flyer identification system provided in the present application. As shown in fig. 2, on the basis of any one of the above embodiments, the flyer identification system provided in the embodiment of the present application is disposed in an intelligent terminal, and the system includes: a binocular fusion module 210, a flyer recognition module 220, and a flyer positioning module 230, wherein:
the binocular fusion module 210 is configured to splice the first image frame and the second image frame to obtain a scene fusion image; the first image frame and the second image frame are respectively obtained by shooting by two cameras with parallel optical axes in the binocular camera;
the flyer identification module 220 is configured to input the scene fusion image into an identification model, and obtain an identified flyer output by the identification model; the identification model is obtained by training the identification flying object type label marked by the sample scene image based on the sample scene image;
a flyer positioning module 230, configured to obtain spatial coordinates of the identified flyer in a phase coordinate system based on two-dimensional image coordinates of the identified flyer in the first image frame and the second image frame, respectively; the phase plane coordinate system is established according to a base line formed between two cameras with parallel optical axes in the binocular camera.
Specifically, binocular fusion module 210, flyer identification module 220, and flyer positioning module 230 are electrically connected in sequence.
The binocular fusion module 210 performs feature extraction on the first image frame and the second image frame acquired by the binocular camera at the same time, and performs registration by using the extracted features, so as to obtain a scene fusion image from which a non-overlapping portion in the two images is removed.
The flyer identification module 220 restores the identification model according to the weight coefficients of the neural network input layer and the hidden layer, and inputs the scene fusion image obtained by the binocular fusion module 210 into the trained identification model, so that the identified flyers of each category identified by the scene fusion image can be obtained.
After the flying object positioning module 230 clearly identifies the flying objects, two-dimensional image coordinates of the center of each identified flying object in the image coordinate system where the first image frame and the second image frame are respectively located are respectively obtained, and based on the two-dimensional image coordinates, the spatial coordinates of the identified flying objects in the phase plane coordinate system are converted, so that the binocular camera can be used for stereoscopic imaging.
Optionally, the flyer positioning module 230 includes a parallax information determining unit and a flyer positioning unit, wherein:
And the parallax information determining unit is used for determining parallax information based on the X-axis coordinate value in the two-dimensional image coordinates of the identification flying object and the depth information of the identification flying object.
And the flying object positioning unit is used for determining the space coordinates of the identified flying object under a phase plane coordinate system based on the two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame and the parallax information.
Optionally, the binocular fusion module 210 includes a feature point extraction unit and a registration unit, wherein:
and the characteristic point extraction unit is used for respectively carrying out characteristic extraction on the first image frame and the second image frame to obtain a first characteristic point set and a second characteristic point set.
And the registration unit is used for matching based on the first characteristic point set and the second characteristic point set to obtain the scene fusion image.
Optionally, the system further comprises a three-dimensional reconstruction module, a coordinate system conversion module, and a real world positioning module, wherein:
the three-dimensional reconstruction module is used for generating a three-dimensional space model of the identified flying object under a camera coordinate system based on the first feature point set, the second feature point set, the first feature depth information and the second feature depth information.
And the coordinate system conversion module is used for acquiring the position coordinates of the identification flying object relative to the camera according to the space coordinates of the identification flying object under the phase plane coordinate system and the three-dimensional space model of the identification flying object under the camera coordinate system.
And the real world positioning module is used for calculating the position coordinates of the identified flying object under the earth coordinate system based on the position coordinates of the identified flying object relative to the camera and the camera gesture data.
The camera coordinate system is established by taking the intelligent terminal as a coordinate origin; the first characteristic depth information comprises depth information corresponding to each characteristic point in the first characteristic point set; the second feature depth information includes depth information corresponding to each feature point in the second feature point set.
Optionally, the three-dimensional reconstruction module includes a point cloud acquisition unit and a three-dimensional reconstruction unit, wherein:
and the point cloud acquisition unit is used for generating sparse point cloud information based on the first characteristic point set, the second characteristic point set, the first characteristic depth information and the second characteristic depth information.
And the three-dimensional reconstruction unit is used for encrypting and smoothing the sparse point cloud information to form the three-dimensional space model of the identified flying object under the camera coordinate system.
Optionally, the recognition model includes a feature extraction layer and a classification layer.
Accordingly, the flyer recognition module 220 includes a feature extraction unit and a classification unit, wherein:
and the feature extraction unit is used for inputting the scene fusion image into the feature extraction layer and obtaining the feature vector output by the feature extraction layer.
And the classifying unit is used for inputting the characteristic vector to the classifying layer and processing the characteristic vector by adopting a multi-head attention mechanism to acquire the identified flying object output by the classifying layer.
Optionally, the system further comprises a calibration module, wherein:
and the calibration module is used for obtaining the internal reference matrix of the binocular camera through calibration so as to calculate the position coordinate of the identified flying object under the earth coordinate system based on the position coordinate of the identified flying object relative to the camera, the camera gesture data and the internal reference matrix of the binocular camera.
The embodiment of the application provides a flyer identification system for executing the flyer identification method, and its implementation mode is consistent with the implementation mode of the flyer identification method provided by the application, and can achieve the same beneficial effects, and the detailed description is omitted here.
After the first image frame and the second image frame which are respectively acquired by the binocular camera are fused, a scene fusion image which only keeps an overlapped part is used as input of an identification model, and an output result is an identification flying object which corresponds to a frame in the image, so that the two-dimensional image coordinates of the identification flying object in the first image frame and the second image frame are utilized to convert the space coordinates of the identification flying object in a phase plane coordinate system to carry out stereoscopic vision reconstruction. Aiming at the intelligent terminal with smaller carrier volume and power consumption, the flying object can be rapidly identified under the condition of no manual intervention, the stereoscopic vision reconstruction work about the identification of the flying object is developed, the identification precision and the efficiency are considered under the condition of low cost, the working intensity of monitoring and the like is further reduced, and the working efficiency is improved.
Fig. 3 is a schematic structural diagram of an intelligent terminal provided in the present application. As shown in fig. 3, on the basis of any one of the above embodiments, the intelligent terminal provided in the embodiments of the present application includes an intelligent terminal body 300 provided with a binocular camera 310, and a vision processor 320 communicatively connected to the binocular camera 310. The vision processor 320, when executing a program, implements the method for identifying flying objects as described in any one of the above.
The intelligent terminal body 300 comprises a mobile phone terminal, augmented reality glasses, mixed reality glasses and augmented reality glasses.
Specifically, the smart terminal uses the smart terminal body 300 as a mounting platform, and a binocular camera 310 and a vision processor 320 which are in communication connection are disposed on the platform.
The vision processor 320 performs feature extraction on the first image frame and the second image frame acquired by the binocular camera 310 at the same time, and performs registration by using the extracted features, so as to obtain a scene fusion image from which a non-overlapping portion in the two images is removed.
The vision processor 320 restores the recognition model according to the weight coefficients of the neural network input layer and the hidden layer, and inputs the obtained scene fusion image into the trained recognition model, so that the recognition flying object of each category recognized by the scene fusion image can be obtained.
After the visual processor 320 clearly identifies the flying objects, two-dimensional image coordinates of the center of each identified flying object in the image coordinate system where the first image frame and the second image frame are respectively located are respectively obtained, and based on the two-dimensional image coordinates, spatial coordinates of the identified flying objects in the phase coordinate system are converted, so that stereoscopic imaging can be performed on the binocular camera.
It is understood that the smart terminal body 300 may be a handheld smart terminal, such as a mobile phone, a smart watch, a tablet computer, etc. The smart terminal body 300 may also be a wearable smart terminal, such as augmented reality glasses, mixed reality glasses, and augmented reality glasses. The embodiment of the present application is not particularly limited thereto.
After the first image frame and the second image frame which are respectively acquired by the binocular camera are fused, a scene fusion image which only keeps an overlapped part is used as input of an identification model, and an output result is an identification flying object which corresponds to a frame in the image, so that the two-dimensional image coordinates of the identification flying object in the first image frame and the second image frame are utilized to convert the space coordinates of the identification flying object in a phase plane coordinate system to carry out stereoscopic vision reconstruction. Aiming at the intelligent terminal with smaller carrier volume and power consumption, the flying object can be rapidly identified under the condition of no manual intervention, the stereoscopic vision reconstruction work about the identification of the flying object is developed, the identification precision and the efficiency are considered under the condition of low cost, the working intensity of monitoring and the like is further reduced, and the working efficiency is improved.
Further, the logic instructions may be embodied in the form of software functional units and stored in a computer readable storage medium when sold or used as a stand alone product. Based on such understanding, the technical solution of the present application may be embodied essentially or in a part contributing to the prior art or in a part of the technical solution, in the form of a software product stored in a storage medium, including several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to perform all or part of the steps of the methods described in the embodiments of the present application. And the aforementioned storage medium includes: a usb disk, a removable hard disk, a Read-only memory (ROM), a random access memory (RAM, randomAccessMemory), a magnetic disk, or an optical disk, or other various media capable of storing program codes.
In another aspect, the present application further provides a computer program product, where the computer program product includes a computer program, where the computer program may be stored on a non-transitory computer readable storage medium, where the computer program, when executed by a processor, is capable of executing the method for identifying a flying object provided by the above methods, and the method is applied to an intelligent terminal, and the method includes:
splicing the first image frame and the second image frame to obtain a scene fusion image; the first image frame and the second image frame are respectively obtained by shooting by two cameras with parallel optical axes in the binocular camera;
inputting the scene fusion image into an identification model to obtain an identification flying object output by the identification model; the identification model is obtained by training the identification flying object type label marked by the sample scene image based on the sample scene image;
acquiring space coordinates of the identified flying object under a phase coordinate system based on two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame respectively; the phase plane coordinate system is established according to a base line formed between two cameras with parallel optical axes in the binocular camera.
In yet another aspect, the present application further provides a non-transitory computer readable storage medium having stored thereon a computer program, which when executed by a processor is implemented to perform the method for identifying a flying object provided by the above methods, and the method is applied to an intelligent terminal, and the method includes:
splicing the first image frame and the second image frame to obtain a scene fusion image; the first image frame and the second image frame are respectively obtained by shooting by two cameras with parallel optical axes in the binocular camera;
inputting the scene fusion image into an identification model to obtain an identification flying object output by the identification model; the identification model is obtained by training the identification flying object type label marked by the sample scene image based on the sample scene image;
acquiring space coordinates of the identified flying object under a phase coordinate system based on two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame respectively; the phase plane coordinate system is established according to a base line formed between two cameras with parallel optical axes in the binocular camera.
The system embodiments described above are merely illustrative, wherein the elements illustrated as separate elements may or may not be physically separate, and the elements shown as elements may or may not be physical elements, may be located in one place, or may be distributed over a plurality of network elements. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of this embodiment. Those of ordinary skill in the art will understand and implement the present invention without undue burden.
From the above description of the embodiments, it will be apparent to those skilled in the art that the embodiments may be implemented by means of software plus necessary general hardware platforms, or of course may be implemented by means of hardware. Based on this understanding, the foregoing technical solution may be embodied essentially or in a part contributing to the prior art in the form of a software product, which may be stored in a computer readable storage medium, such as ROM/RAM, a magnetic disk, an optical disk, etc., including several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to execute the method described in the respective embodiments or some parts of the embodiments.
Finally, it should be noted that: the above embodiments are only for illustrating the technical solution of the present application, and are not limiting thereof; although the present application has been described in detail with reference to the foregoing embodiments, it should be understood by those of ordinary skill in the art that: the technical scheme described in the foregoing embodiments can be modified or some technical features thereof can be replaced by equivalents; such modifications and substitutions do not depart from the spirit and scope of the corresponding technical solutions.

Claims (10)

1. The flyer identification method is characterized by being applied to an intelligent terminal and comprising the following steps of:
splicing the first image frame and the second image frame to obtain a scene fusion image; the first image frame and the second image frame are respectively obtained by shooting by two cameras with parallel optical axes in the binocular camera;
inputting the scene fusion image into an identification model to obtain an identification flying object output by the identification model; the identification model is obtained by training the identification flying object type label marked by the sample scene image based on the sample scene image;
acquiring space coordinates of the identified flying object under a phase coordinate system based on two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame respectively; the phase plane coordinate system is established according to a base line formed between two cameras with parallel optical axes in the binocular camera.
2. The method of claim 1, wherein acquiring spatial coordinates of the identified flying object in a phase plane coordinate system based on two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame, respectively, comprises:
Determining parallax information based on the X-axis coordinate value in the two-dimensional image coordinates of the identified flying object and the depth information of the identified flying object;
and determining the space coordinates of the identified flying object under a phase plane coordinate system based on the two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame respectively and the parallax information.
3. The method of claim 1, wherein the stitching the first image frame and the second image frame to obtain the scene fusion image comprises:
respectively extracting features of the first image frame and the second image frame to obtain a first feature point set and a second feature point set;
and matching based on the first characteristic point set and the second characteristic point set to obtain the scene fusion image.
4. A method of identifying an object as claimed in claim 3, further comprising, after said obtaining the spatial coordinates of the identified object in the phase plane coordinate system:
generating a three-dimensional space model of the identified flying object under a camera coordinate system based on the first feature point set and the second feature point set, and the first feature depth information and the second feature depth information;
Acquiring the position coordinates of the identified flying object relative to a camera according to the space coordinates of the identified flying object under the phase plane coordinate system and the three-dimensional space model of the identified flying object under the camera coordinate system;
based on the position coordinates of the identified flying object relative to the camera and the camera gesture data, calculating the position coordinates of the identified flying object under an earth coordinate system;
the camera coordinate system is established by taking the intelligent terminal as a coordinate origin; the first characteristic depth information comprises depth information corresponding to each characteristic point in the first characteristic point set; the second feature depth information includes depth information corresponding to each feature point in the second feature point set.
5. The method of claim 4, wherein generating the three-dimensional model of the identified flying object in the camera coordinate system based on the first set of feature points and the second set of feature points, and the first feature depth information and the second feature depth information, comprises:
generating sparse point cloud information based on the first feature point set and the second feature point set, and the first feature depth information and the second feature depth information;
And encrypting and smoothing the sparse point cloud information to form the three-dimensional space model of the identified flying object under the camera coordinate system.
6. The method of claim 1, wherein the recognition model includes a feature extraction layer and a classification layer;
correspondingly, the step of inputting the scene fusion image into the recognition model to acquire the recognized flying object output by the recognition model comprises the following steps:
inputting the scene fusion image into the feature extraction layer to obtain a feature vector output by the feature extraction layer;
and inputting the feature vector to the classification layer, and processing by adopting a multi-head attention mechanism to obtain the identified flying object output by the classification layer.
7. The method for identifying a flying object according to any one of claims 1 to 6, further comprising, before the first image frame and the second image frame are spliced to obtain a scene fusion image:
and obtaining an internal reference matrix of the binocular camera through calibration so as to calculate the position coordinate of the identified flying object under the earth coordinate system based on the position coordinate of the identified flying object relative to the camera, the camera gesture data and the internal reference matrix of the binocular camera.
8. The utility model provides a flyer identification system which characterized in that sets up in intelligent terminal, includes:
the binocular fusion module is used for splicing the first image frame and the second image frame to obtain a scene fusion image; the first image frame and the second image frame are respectively obtained by shooting by two cameras with parallel optical axes in the binocular camera;
the flyer identification module is used for inputting the scene fusion image into an identification model to acquire an identified flyer output by the identification model; the recognition model is obtained by training a recognition flyer category label marked by a sample scene image;
the flying object positioning module is used for acquiring space coordinates of the identified flying object under a phase plane coordinate system based on two-dimensional image coordinates of the identified flying object in the first image frame and the second image frame respectively; the phase plane coordinate system is established according to a base line formed between two cameras with parallel optical axes in the binocular camera.
9. The intelligent terminal is characterized by comprising an intelligent terminal body provided with a binocular camera and a vision processor in communication connection with the binocular camera; the vision processor implementing the method for identifying a flying object according to any one of claims 1 to 7 when executing a program;
The intelligent terminal body comprises a mobile phone terminal, augmented reality glasses, mixed reality glasses and augmented reality glasses.
10. A non-transitory computer readable storage medium, on which a computer program is stored, characterized in that the computer program, when executed by a processor, implements the method of identifying flying objects according to any one of claims 1 to 7.
CN202410001150.8A 2024-01-02 2024-01-02 Flying object identification method, system, intelligent terminal and computer readable storage medium Pending CN117495698A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202410001150.8A CN117495698A (en) 2024-01-02 2024-01-02 Flying object identification method, system, intelligent terminal and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202410001150.8A CN117495698A (en) 2024-01-02 2024-01-02 Flying object identification method, system, intelligent terminal and computer readable storage medium

Publications (1)

Publication Number Publication Date
CN117495698A true CN117495698A (en) 2024-02-02

Family

ID=89681484

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202410001150.8A Pending CN117495698A (en) 2024-01-02 2024-01-02 Flying object identification method, system, intelligent terminal and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN117495698A (en)

Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106570908A (en) * 2016-11-14 2017-04-19 中北大学 Stereoscopic vision apparatus for testing scattering area of mother-son flyer
US20190204084A1 (en) * 2017-09-29 2019-07-04 Goertek Inc. Binocular vision localization method, device and system
CN110390719A (en) * 2019-05-07 2019-10-29 香港光云科技有限公司 Based on flight time point cloud reconstructing apparatus
CN111856448A (en) * 2020-07-02 2020-10-30 山东省科学院海洋仪器仪表研究所 Marine obstacle identification method and system based on binocular vision and radar
CN113160298A (en) * 2021-03-31 2021-07-23 奥比中光科技集团股份有限公司 Depth truth value acquisition method, device and system and depth camera
CN113269820A (en) * 2021-05-26 2021-08-17 北京地平线信息技术有限公司 Method and device for generating space geometric information estimation model
CN113296133A (en) * 2020-04-01 2021-08-24 易通共享技术(广州)有限公司 Device and method for realizing position calibration based on binocular vision measurement and high-precision positioning fusion technology
CN113524194A (en) * 2021-04-28 2021-10-22 重庆理工大学 Target grabbing method of robot vision grabbing system based on multi-mode feature deep learning
CN113850195A (en) * 2021-09-27 2021-12-28 杭州东信北邮信息技术有限公司 AI intelligent object identification method based on 3D vision
CN113954085A (en) * 2021-09-08 2022-01-21 重庆大学 Intelligent positioning and control method of welding robot based on binocular vision and linear laser sensing data fusion
WO2022052366A1 (en) * 2020-09-08 2022-03-17 奥比中光科技集团股份有限公司 Fused depth measurement method and measurement device
CN114596382A (en) * 2022-02-19 2022-06-07 复旦大学 Binocular vision SLAM method and system based on panoramic camera
CN115115768A (en) * 2021-03-18 2022-09-27 上海宝信软件股份有限公司 Object coordinate recognition system, method, device and medium based on stereoscopic vision
CN115497077A (en) * 2022-10-24 2022-12-20 广西柳工机械股份有限公司 Carriage attitude recognition system, carriage attitude recognition method, electronic device and storage medium
CN115546741A (en) * 2022-09-29 2022-12-30 华南理工大学 Binocular vision and laser radar unmanned ship marine environment obstacle identification method
CN115880368A (en) * 2022-09-30 2023-03-31 国网湖南省电力有限公司 Method and system for detecting obstacle of power grid inspection unmanned aerial vehicle and storage medium
CN116205947A (en) * 2023-01-03 2023-06-02 哈尔滨工业大学 Binocular-inertial fusion pose estimation method based on camera motion state, electronic equipment and storage medium
CN117115271A (en) * 2023-08-31 2023-11-24 山东大学 Binocular camera external parameter self-calibration method and system in unmanned aerial vehicle flight process

Patent Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106570908A (en) * 2016-11-14 2017-04-19 中北大学 Stereoscopic vision apparatus for testing scattering area of mother-son flyer
US20190204084A1 (en) * 2017-09-29 2019-07-04 Goertek Inc. Binocular vision localization method, device and system
CN110390719A (en) * 2019-05-07 2019-10-29 香港光云科技有限公司 Based on flight time point cloud reconstructing apparatus
CN113296133A (en) * 2020-04-01 2021-08-24 易通共享技术(广州)有限公司 Device and method for realizing position calibration based on binocular vision measurement and high-precision positioning fusion technology
CN111856448A (en) * 2020-07-02 2020-10-30 山东省科学院海洋仪器仪表研究所 Marine obstacle identification method and system based on binocular vision and radar
WO2022052366A1 (en) * 2020-09-08 2022-03-17 奥比中光科技集团股份有限公司 Fused depth measurement method and measurement device
CN115115768A (en) * 2021-03-18 2022-09-27 上海宝信软件股份有限公司 Object coordinate recognition system, method, device and medium based on stereoscopic vision
CN113160298A (en) * 2021-03-31 2021-07-23 奥比中光科技集团股份有限公司 Depth truth value acquisition method, device and system and depth camera
CN113524194A (en) * 2021-04-28 2021-10-22 重庆理工大学 Target grabbing method of robot vision grabbing system based on multi-mode feature deep learning
CN113269820A (en) * 2021-05-26 2021-08-17 北京地平线信息技术有限公司 Method and device for generating space geometric information estimation model
WO2022247414A1 (en) * 2021-05-26 2022-12-01 北京地平线信息技术有限公司 Method and apparatus for generating space geometry information estimation model
CN113954085A (en) * 2021-09-08 2022-01-21 重庆大学 Intelligent positioning and control method of welding robot based on binocular vision and linear laser sensing data fusion
CN113850195A (en) * 2021-09-27 2021-12-28 杭州东信北邮信息技术有限公司 AI intelligent object identification method based on 3D vision
CN114596382A (en) * 2022-02-19 2022-06-07 复旦大学 Binocular vision SLAM method and system based on panoramic camera
CN115546741A (en) * 2022-09-29 2022-12-30 华南理工大学 Binocular vision and laser radar unmanned ship marine environment obstacle identification method
CN115880368A (en) * 2022-09-30 2023-03-31 国网湖南省电力有限公司 Method and system for detecting obstacle of power grid inspection unmanned aerial vehicle and storage medium
CN115497077A (en) * 2022-10-24 2022-12-20 广西柳工机械股份有限公司 Carriage attitude recognition system, carriage attitude recognition method, electronic device and storage medium
CN116205947A (en) * 2023-01-03 2023-06-02 哈尔滨工业大学 Binocular-inertial fusion pose estimation method based on camera motion state, electronic equipment and storage medium
CN117115271A (en) * 2023-08-31 2023-11-24 山东大学 Binocular camera external parameter self-calibration method and system in unmanned aerial vehicle flight process

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
4SNJ_VISION263C: "视觉识别项目:双目摄像头进行空间定位技术详解", pages 1 - 16, Retrieved from the Internet <URL:https://www.elecfans.com/kongzhijishu/jiqishijue/2032522.html> *
梁乐: ""基于双目立体视觉的不规则物体体积测量方法研究", 中国优秀硕士学位论文全文数据库信息科技辑, no. 08, 15 August 2019 (2019-08-15), pages 5 - 20 *
董保磊 等;: "基于双目视觉的无人机悬停精度测量研究", 计算机工程与应用, no. 04, 14 September 2017 (2017-09-14) *
邾继贵 等;: "光学坐标测量系统中的关键技术", 红外与激光工程, no. 03, 25 June 2007 (2007-06-25), pages 296 - 299 *
郭子兴 等;: "基于双目视觉的增强现实系统设计与实现", 电子设计工程, no. 23, 5 December 2018 (2018-12-05) *
高少薄: "基于双目视觉的集装箱位姿测量研究", 《中国优秀硕士学位论文全文数据库信息科技辑, no. 02, 15 February 2023 (2023-02-15), pages 10 - 24 *

Similar Documents

Publication Publication Date Title
CN110568447B (en) Visual positioning method, device and computer readable medium
US9928656B2 (en) Markerless multi-user, multi-object augmented reality on mobile devices
CN106529538A (en) Method and device for positioning aircraft
CN111623765B (en) Indoor positioning method and system based on multi-mode data
CN104156998A (en) Implementation method and system based on fusion of virtual image contents and real scene
CN113256718B (en) Positioning method and device, equipment and storage medium
AliAkbarpour et al. Fast structure from motion for sequential and wide area motion imagery
CN115035235A (en) Three-dimensional reconstruction method and device
CN112528974A (en) Distance measuring method and device, electronic equipment and readable storage medium
CN108010122B (en) Method and system for reconstructing and measuring three-dimensional model of human body
CN114627491A (en) Single three-dimensional attitude estimation method based on polar line convergence
CN112150518A (en) Attention mechanism-based image stereo matching method and binocular device
CN110443228B (en) Pedestrian matching method and device, electronic equipment and storage medium
CN110012236A (en) A kind of information processing method, device, equipment and computer storage medium
CN113902802A (en) Visual positioning method and related device, electronic equipment and storage medium
CN113361365A (en) Positioning method and device, equipment and storage medium
CN108335329B (en) Position detection method and device applied to aircraft and aircraft
KR101673144B1 (en) Stereoscopic image registration method based on a partial linear method
Price et al. Augmenting crowd-sourced 3d reconstructions using semantic detections
CN116051736A (en) Three-dimensional reconstruction method, device, edge equipment and storage medium
CN117495698A (en) Flying object identification method, system, intelligent terminal and computer readable storage medium
Workman et al. Augmenting depth estimation with geospatial context
CN114494857A (en) Indoor target object identification and distance measurement method based on machine vision
CN115482285A (en) Image alignment method, device, equipment and storage medium
CN113281770A (en) Coordinate system relation obtaining method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination