CN115514885A - Monocular and binocular fusion-based remote augmented reality follow-up perception system and method - Google Patents

Monocular and binocular fusion-based remote augmented reality follow-up perception system and method Download PDF

Info

Publication number
CN115514885A
CN115514885A CN202211037134.1A CN202211037134A CN115514885A CN 115514885 A CN115514885 A CN 115514885A CN 202211037134 A CN202211037134 A CN 202211037134A CN 115514885 A CN115514885 A CN 115514885A
Authority
CN
China
Prior art keywords
bucket
follow
information
monocular
augmented reality
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202211037134.1A
Other languages
Chinese (zh)
Other versions
CN115514885B (en
Inventor
丁伟利
李健
华长春
魏饶
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Yanshan University
Original Assignee
Yanshan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Yanshan University filed Critical Yanshan University
Priority to CN202211037134.1A priority Critical patent/CN115514885B/en
Publication of CN115514885A publication Critical patent/CN115514885A/en
Application granted granted Critical
Publication of CN115514885B publication Critical patent/CN115514885B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/012Head tracking input arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/332Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/366Image reproducers using viewer tracking

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The invention discloses a remote augmented reality follow-up sensing system and a remote augmented reality follow-up sensing method based on monocular and binocular fusion, which belong to the field of intelligent engineering machinery, wherein the system comprises a monocular and binocular fusion follow-up intelligent sensing module, a server-side intelligent processing module and a user-side augmented reality module, wherein the monocular and binocular fusion follow-up intelligent sensing module is used for acquiring RGB information and depth data of a construction scene and sending the RGB information and the depth data to the edge end of a server through a wireless transmission technology; the server-side intelligent processing module is used for executing calculations required by bucket attitude estimation, bucket tip positioning, accurate environment perception and augmented reality; the user-side augmented reality module comprises three-dimensional display equipment and an operation console, the three-dimensional display equipment is used for displaying the three-dimensional information fusion image processed by the server-side algorithm processing module, and the operation console is used for controlling engineering machinery operation on a construction site. The invention is based on the remote follow-up intelligent sensing technology, and can solve the problems of absence of telepresence and distance sense of an operator and the like.

Description

Monocular and binocular fusion-based remote augmented reality follow-up perception system and method
Technical Field
The invention relates to the field of intelligent engineering machinery, in particular to a monocular and binocular fusion-based remote augmented reality follow-up sensing system and method.
Background
In domestic and international markets, the requirements of engineering machinery users on equipment are continuously diversified, and the requirements of people on operation comfort are higher and higher. Especially in hard, high-risk and repetitive operation environments, because the health of operators cannot be guaranteed, skilled operators are seriously in short supply, and the recruitment of workers in construction enterprises is more and more difficult. In these environments and accident disasters, a large number of engineering machines such as loaders and excavators are often required to perform emergency rescue such as on-site cleaning and road restoration, and people are increasingly eager for obtaining comfortable working environment and equivalent working efficiency by remotely and intelligently remotely controlling the engineering machines. However, the environment sensing of the teleoperation system of the current engineering machinery generally adopts a visible light camera or a laser radar with a fixed position relative to the machine body, and limits the field range of intelligent sensing to a certain extent.
The invention with the application number of 201810268923.3 discloses a stereoscopic vision follow-up system applied to disaster area searching, which obtains disaster area field audio and video in real time by wearing VR glasses and Bluetooth earphones by an operator, and controls the synchronous motion of an unmanned aerial vehicle camera by using the head posture; the invention with the application number of 202010882933.3 discloses a humanoid binocular follow-up virtual reality system suitable for teleoperation of a robot, wherein a binocular camera is placed on a two-dimensional platform, and the binocular camera is controlled by a follow-up mechanism to synchronously pitch and rotate along with the head movement of an operator, so that the visual angle of the binocular camera is changed. Although the technology provides an immersive stereoscopic impression for operators, the technology has the following defects: only binocular images are directly provided for operators, and the problems of absence of presence and distance are still caused.
Disclosure of Invention
The system is based on a remote follow-up intelligent sensing technology, can provide important information such as real-time posture estimation of a bucket and bucket tip positioning of the bucket for an operator, and solves the problems of absence of telepresence and distance sense of the operator.
In order to solve the technical problems, the technical scheme adopted by the invention is as follows:
a remote augmented reality follow-up sensing system based on monocular and binocular fusion comprises an edge end monocular and binocular fusion follow-up intelligent sensing module, a server end intelligent processing module and a user end augmented reality module;
the edge end single-binocular fusion follow-up intelligent sensing module is used for acquiring RGB information and depth data of a construction scene based on a single-binocular fusion method and sending the RGB information and the depth data to the server end through wireless transmission;
the server-side intelligent processing module is used for executing calculations required by bucket attitude estimation, bucket tip positioning, accurate environment perception and augmented reality;
the user-side augmented reality module comprises a three-dimensional display device and a control console, wherein the three-dimensional display device is used for displaying the three-dimensional information fusion image processed by the server-side algorithm processing module, and the control console is used for controlling the operation of the engineering machinery on a construction site.
The technical scheme of the invention is further improved as follows: the edge end single-binocular fusion follow-up intelligent sensing module comprises a single-binocular vision sensor, an edge end AI processor and a follow-up holder; the edge terminal AI processor is used for realizing the fusion perception of single and double eye RGB information, depth data, camera pose information and key target information, reading the head posture of an operator wearing video glasses at a user terminal, and further controlling a follow-up cradle head carrying a single and double eye vision sensor to quickly synchronize the head posture of the operator by a direct current brushless motor control method, so that a follow-up effect is achieved.
The technical scheme of the invention is further improved as follows: the edge terminal AI processor is connected with the monocular and binocular vision sensor through a USB data line, and reads monocular RGB information, binocular gray scale information and camera pose information in real time; the binocular gray scale information adopts a stereo matching algorithm to recover a depth map, 2D points on the depth map are converted into 3D points under a world coordinate system through internal and external parameters of a binocular depth camera, and then the 3D points under the world coordinate system are projected onto an RGB image through the internal and external parameters of a monocular RGB camera, so that single and binocular information fusion is realized; and detecting the information of the key target in real time based on a target detection algorithm, and sending the position of the key target to a server-side intelligent processing module for positioning the bucket tip.
The technical scheme of the invention is further improved as follows: the intelligent processing module at the server side adopts an efficient sparse region template matching method and a real-time lightweight deep learning network based bucket attitude estimation and positioning algorithm to track the state of the bucket in real time, and the accurate environment perception algorithm adopts a monocular vision SLAM algorithm, so that necessary environment map information and engineering machinery attitude information can be provided for safe operation.
The technical scheme of the invention is further improved as follows: the three-dimensional display equipment can accurately capture the head posture of an operator while displaying the fused image processed by the server end, and sends the head posture to the server end in real time, so that the head posture is read by the edge end AI processor and the follow-up holder is controlled to follow; the control console can also be used for providing a real control environment for an operator, and can achieve an effect of being personally on the scene by matching with the video glasses.
A remote augmented reality follow-up perception method based on monocular and binocular fusion comprises the following steps:
step 1, placing an edge end single-binocular fusion follow-up intelligent sensing module in a cab of engineering machinery, and after ensuring that glass in front of the cab is not shielded, turning on a power supply of an edge end processor to enable the edge end processor to be in a waiting state, and waiting for establishing communication connection with a server end intelligent processing module;
step 2, turning on a power supply of the server side to enable the server side to be in a monitoring state, and waiting for establishing communication connection with the edge end single-binocular fusion follow-up intelligent sensing module and the user side augmented reality module;
step 3, an operator enters the console, wears the video glasses and starts remote control operation after the three-dimensional display equipment has a construction interface;
step 4, the edge processor reads the head posture data of the operator, controls the follow-up cradle head to update the posture in real time, and simultaneously sends the RGB information and the depth data of the construction scene to the server end through a wireless transmission technology;
step 5, the server-side intelligent processing module carries out attitude estimation and bucket tip positioning of the bucket and environment map information construction through the received RGB information and depth data, and finally sends a fused image fused with the bucket attitude information, bucket tip position information and actual distance information of surrounding objects such as the bucket tip and a dump truck to a user side for display;
step 6, the user side operator remotely controls the engineering machinery operation in a manner of being matched with the console through the fused image and the on-site three-dimensional information displayed by the video glasses, and meanwhile, the video glasses capture the head posture of the operator in real time and send the head posture to the server side to wait for the edge side processor to read;
and 7, repeating the step 4 to the step 6.
A bucket attitude tracking method based on a sparse region uses a remote augmented reality follow-up perception system based on single and binocular fusion, and comprises the following steps:
s1, placing a bucket under natural illumination, avoiding reflective objects around the bucket, taking 30 pictures by using photographing equipment to surround the bucket for a circle, and enabling the bucket to be located at the center of an image during photographing;
s2, opening RealityCapture software, and generating a bucket three-dimensional model by utilizing 30 bucket photos, wherein the three-dimensional model is completely the same as the real bucket in proportion;
s3, placing virtual cameras at 2562 different positions around the three-dimensional model to render the three-dimensional model, acquiring sparse contour points of the bucket in the current posture by using a rendering map, back-projecting the contour points to a coordinate system of the bucket three-dimensional model for storage, and simultaneously storing normal vectors of the contour points and direction vectors of the current posture; 2562 template views are finally generated;
s4, giving an initial posture of the bucket, multiplying the direction vectors of all the template views by the initial posture, and finding out the template view consistent with the initial posture; projecting the contour point of the template view onto a current real image, wherein the front 18 pixels of the contour point along the normal direction are designated as bucket pixels, the rear 18 pixels are designated as background pixels, and the bucket and the background are segmented to obtain the real contour of the bucket;
and S5, estimating the real posture of the bucket by using the distance between the model contour point and the real contour point along the normal direction, and further realizing the bucket tracking.
Due to the adoption of the technical scheme, the invention has the technical progress that:
1. due to the adoption of the monocular and binocular follow-up intelligent sensing technology, the invention realizes that the operator gets rid of the limitation of fixed visual angle of the traditional teleoperation under the condition of over-sight distance, and monocular RGB information and binocular depth data of different visual angles of a construction site can be freely acquired through the change of the head posture.
2. Due to the adoption of the efficient and real-time bucket tracking and bucket tip positioning algorithm, the posture of the bucket and the spatial position of the bucket tip can be captured in real time through images, so that the real-time posture information and the position information of the bucket are provided for operators by matching with an augmented reality technology, and the limitation that the distance sense cannot be provided for the operators by the traditional teleoperation is overcome.
3. Due to the adoption of the SLAM + YOLO environment perception algorithm, the local coordinates of the bucket tip of the bucket can be transformed to the starting point coordinate system of the camera in real time, so that the global environment perception is provided, and a foundation is provided for the efficient operation of the engineering machinery.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings required to be used in the description of the embodiments or the prior art will be briefly introduced below, and it is obvious that the drawings in the following description are some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts;
fig. 1 is a schematic diagram of an overall structure of a remote augmented reality follow-up sensing system based on monocular and binocular fusion in an embodiment of the present invention;
fig. 2 is a schematic diagram of a general structure of a remote augmented reality follow-up sensing system based on monocular and binocular fusion in the embodiment of the present invention;
fig. 3 is a structure diagram of a servo pan-tilt of a remote augmented reality servo perception system based on monocular and binocular fusion in the embodiment of the present invention;
FIG. 4 is a schematic diagram of monocular and binocular fusion of the remote augmented reality follow-up sensing system based on monocular and binocular fusion in the embodiment of the present invention;
FIG. 5 is a schematic diagram of bucket tracking of a remote augmented reality follow-up sensing system based on monocular and binocular fusion according to an embodiment of the present invention;
FIG. 6 is a schematic diagram illustrating positioning of a bucket tip of a bucket of a remote augmented reality follow-up sensing system based on monocular and binocular fusion in the embodiment of the present invention;
the device comprises an edge terminal AI processor 1, a monocular and binocular vision sensor 2, a 3, y-axis brushless DC motor, a 4, x-axis brushless DC motor, a 5, RGB camera, a 6, left eye depth camera and a 7, right eye depth camera.
Detailed Description
It should be noted that the terms "comprises" and "comprising," and any variations thereof, in the description and claims of the present invention and the above-described drawings, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
The invention is described in further detail below with reference to the following figures and examples:
as shown in fig. 1 and 2, a monocular and binocular fusion based remote augmented reality follow-up sensing system includes an edge end monocular and binocular fusion follow-up intelligent sensing module, a server end intelligent processing module and a user end augmented reality module;
the edge end single-binocular fusion follow-up intelligent sensing module is placed in a cab of the engineering machinery, is used for acquiring RGB information and depth data of a construction scene and sending the RGB information and the depth data to the server end through a wireless transmission technology;
the server-side intelligent processing module is used for executing calculation required by bucket attitude estimation, bucket tip positioning, accurate environment perception and augmented reality;
the user side augmented reality module comprises video glasses and a console, the video glasses are used for displaying the three-dimensional information fusion image processed by the server side algorithm processing module, and the console is used for controlling engineering machinery operation of a construction site.
As shown in fig. 3, the servo-actuated tripod head integrates a monocular and binocular vision sensor 2, an edge AI processor 1, an x-axis dc brushless motor 4 and a y-axis dc brushless motor 3, adopts the self-design of solidworks software, has two degrees of freedom of the x-axis and the y-axis, and enables the dc motor to have small torque fluctuation, high efficiency, small noise and fast dynamic response through a motor control technology (FOC motor control technology), thereby being capable of rapidly synchronizing the head posture of an operator.
As shown in fig. 4, the monocular and binocular vision sensor 2 has three cameras, which are an RGB camera 5, a left eye depth camera 6 and a right eye depth camera 7, the RGB camera 5 can collect RGB information including a bucket, and sends the RGB information to the server-side intelligent processing module for bucket tracking through wireless transmission, the left eye depth camera 6 and the right eye depth camera 7 can collect gray scale information including the bucket, the resolution of the RGB image is kept consistent through upsampling, depth recovery is performed based on the right eye image by using an SGBM binocular stereo matching algorithm, and finally, a depth image pixel point is re-projected onto the RGB image through a coordinate transformation relation between the right eye depth camera 7 and the RGB camera 5 for monocular and binocular information fusion, so that bucket attitude information can be transformed into an RGB coordinate system, and the bucket point can be conveniently located subsequently.
As shown in fig. 5, a bucket attitude estimation algorithm deployed by the server-side intelligent processing module can generate template views of each attitude by using a three-dimensional model of the bucket, and then match the template views with a real view of the bucket returned by the servo pan-tilt to obtain accurate estimation of the bucket attitude, and then superimpose and display a rendering image of the current attitude and the real image to visualize a tracking result.
As shown in fig. 6 (a), the bucket tip positioning algorithm deployed by the server-side intelligent processing module can obtain two-dimensional image plane coordinates of the bucket tip by using a two-dimensional image rendered by the bucket three-dimensional model, and then obtain three-dimensional coordinates of the bucket tip in the three-dimensional model coordinate system through internal and external parameters of the virtual camera. As shown in fig. 6 (b), the bucket tip coordinates in the three-dimensional model coordinate system may be transformed to the camera coordinate system through the camera pose, the position of the dump truck is detected through the YOLOv4 target detection algorithm deployed by the edge AI processor 1, the three-dimensional coordinates of the truck center point in the camera coordinate system are obtained by matching with the depth data, the euclidean distance between the bucket tip and the truck center point is further calculated, the information of the relative distance between the bucket tip and the truck is provided for the operator, and when the relative distance is smaller than the set threshold, an alarm may be issued to the operator to avoid collision. As shown in (c) in fig. 6, by using the SLAM + YOLO algorithm deployed on the server side, the points outside the box represent the detected static feature points, the points inside the box represent the bucket and the dynamic feature points around the bucket, and the transformation relationship of the current frame with respect to the first frame is calculated by using the static feature points of each frame, so that the tip coordinates at each time are converted into the coordinates in the camera coordinate system at the initial time, that is, the coordinates in the global coordinate system. The global bucket coordinate can better estimate the construction state of the current engineering machine.
A remote augmented reality follow-up perception method based on monocular and binocular fusion comprises the following steps:
step 1, placing an edge end single-binocular fusion follow-up intelligent sensing module in a cab of engineering machinery, and after ensuring that glass in front of the cab is not shielded, turning on a power supply of an edge end processor to enable the edge end processor to be in a waiting state, and waiting for establishing communication connection with a server end intelligent processing module;
step 2, turning on a power supply of a server end to enable the server end to be in a monitoring state, and waiting for establishing communication connection with the edge end single-binocular fusion follow-up intelligent sensing module and the user end augmented reality module;
step 3, an operator enters the console, wears the video glasses and starts remote control operation after the three-dimensional display equipment has a construction interface;
step 4, the edge processor reads the head posture data of the operator, controls the follow-up cradle head to update the posture in real time, and simultaneously sends the RGB information and the depth data of the construction scene to the server end through a wireless transmission technology;
step 5, the server-side intelligent processing module carries out attitude estimation and bucket tip positioning of the bucket and environment map information construction through the received RGB information and depth data, and finally sends a fused image fused with the bucket attitude information, bucket tip position information and actual distance information of surrounding objects such as the bucket tip and a dump truck to a user side for display;
step 6, the user side operator remotely controls the engineering machinery operation in a manner of being matched with the console through the fused image and the on-site three-dimensional information displayed by the video glasses, and meanwhile, the video glasses capture the head posture of the operator in real time and send the head posture to the server side to wait for the edge side processor to read;
and 7, repeating the step 4 to the step 6.
Specifically, when the engineering machinery carries out remote control operation, an operator is positioned on the control console and wears the video glasses. The image displayed by the video glasses is a fused image processed and completed by the server-side intelligent processing module, the fused image comprises attitude information of the bucket, distance information between the bucket tip and the servo pan-tilt, distance information between the bucket tip and the dump truck and global environment information of the bucket, and an operator can judge the construction state of a construction site by using the information and control the operation of engineering machinery of the construction site through the console. The follow-up cradle head placed in the engineering machinery cab of the construction site can freely rotate according to the head posture of an operator, on-site RGB images and depth images are collected in real time, then the images are sent to the server-side intelligent processing module through a wireless transmission technology to be subjected to posture estimation, bucket tip positioning, environment perception and image fusion based on the augmented reality technology, and finally the server side sends the fused images to video glasses to be displayed.
The beyond-the-horizon remote augmented reality follow-up intelligent sensing system and method based on monocular and binocular fusion can be used in the existing practical intelligent engineering machinery teleoperation system, an intelligent follow-up cradle head can replace an unmanned aerial vehicle to sense the environment, a bucket tracking method based on vision can replace a tracking method based on an IMU sensor, video glasses can replace a large-screen two-dimensional display interface, and the telepresence sense of being personally on the scene is provided for teleoperators.
The invention also provides a bucket attitude tracking method based on the sparse region, which uses a remote augmented reality follow-up perception system based on monocular and binocular fusion, and the tracking method realizes functions through the following steps:
s1, placing the bucket under natural illumination, and avoiding reflective objects around the bucket as much as possible. 30 pictures are taken with any photographing device around the bucket, with the bucket as centered as possible in the image.
And S2, opening RealityCapture software, and generating a bucket three-dimensional model by using 30 bucket photos, wherein the three-dimensional model is completely the same as a real bucket in proportion.
And S3, placing 2562 virtual cameras at different positions around the three-dimensional model to render the three-dimensional model, acquiring sparse contour points of the bucket in the current posture by using a rendering graph, back-projecting the contour points to a coordinate system of the bucket three-dimensional model for storage, and simultaneously storing normal vectors of the contour points and direction vectors of the current posture. Finally 2562 template views can be generated.
And S4, giving the initial posture of the bucket, multiplying the direction vectors of all the template views by the initial posture, and finding out the template view closest to the initial posture. And projecting the contour point of the template view onto the current real image, wherein the front 18 pixels of the contour point along the normal direction are designated as bucket pixels, and the rear 18 pixels are designated as background pixels, so that the bucket and the background can be segmented, and the real contour of the bucket is obtained.
And S5, estimating the real posture of the bucket by using the distance between the model contour point and the real contour point along the normal direction, and further realizing the bucket tracking.
Finally, it should be noted that: the above embodiments are only used to illustrate the technical solution of the present invention, and not to limit the same; while the invention has been described in detail and with reference to the foregoing embodiments, it will be understood by those skilled in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some or all of the technical features may be equivalently replaced; and these modifications or substitutions do not depart from the spirit of the corresponding technical solutions of the embodiments of the present invention.

Claims (7)

1. The utility model provides a long-range augmented reality follow-up perception system based on monocular and binocular fusion which characterized in that: the system comprises an edge end single-binocular fusion follow-up intelligent sensing module, a server end intelligent processing module and a user end augmented reality module;
the edge end single-binocular fusion follow-up intelligent sensing module is used for acquiring RGB information and depth data of a construction scene based on a single-binocular fusion method and sending the RGB information and the depth data to the server end through wireless transmission;
the server-side intelligent processing module is used for executing calculation required by bucket attitude estimation, bucket tip positioning, accurate environment perception and augmented reality;
the user-side augmented reality module comprises a three-dimensional display device and a control console, wherein the three-dimensional display device is used for displaying the three-dimensional information fusion image processed by the server-side algorithm processing module, and the control console is used for controlling the operation of the engineering machinery on a construction site.
2. The monocular and binocular fusion based remote augmented reality follow-up perception system according to claim 1, wherein: the edge end single-binocular fusion follow-up intelligent sensing module comprises a single-binocular vision sensor, an edge end AI processor and a follow-up holder; the edge terminal AI processor is used for realizing the fusion perception of monocular and binocular RGB information, depth data, camera pose information and key target information, reading the head posture of an operator wearing the video glasses at a user terminal, and controlling a follow-up pan-tilt carrying monocular and binocular vision sensors to quickly synchronize the head posture of the operator by a direct current brushless motor control method, so that a follow-up effect is achieved.
3. The monocular and binocular fusion based remote augmented reality follow-up perception system according to claim 2, wherein: the edge terminal AI processor is connected with the monocular and binocular vision sensor through a USB data line, and reads monocular RGB information, binocular gray scale information and camera pose information in real time; the binocular gray scale information adopts a stereo matching algorithm to recover a depth map, 2D points on the depth map are converted into 3D points under a world coordinate system through internal and external parameters of a binocular depth camera, and then the 3D points under the world coordinate system are projected onto an RGB image through the internal and external parameters of a monocular RGB camera, so that single and binocular information fusion is realized; and detecting the information of the key target in real time based on a target detection algorithm, and sending the position of the key target to a server-side intelligent processing module for positioning the bucket tip.
4. The monocular and binocular fusion based remote augmented reality follow-up perception system according to claim 1, wherein: the intelligent processing module at the server side adopts an efficient sparse region template matching method and a real-time lightweight deep learning network based bucket attitude estimation and positioning algorithm to track the state of the bucket in real time, and the accurate environment perception algorithm adopts a monocular vision SLAM algorithm, so that necessary environment map information and engineering machinery attitude information can be provided for safe operation.
5. The monocular and binocular fusion based remote augmented reality follow-up perception system according to claim 1, wherein: the three-dimensional display equipment can accurately capture the head posture of an operator while displaying the fused image processed by the server, and sends the head posture to the server in real time, so that the head posture is read by the edge terminal AI processor to control the follow-up cradle head to follow; the control console can also be used for providing a real control environment for an operator, and can achieve the effect of being personally on the scene by matching with video glasses.
6. A perception method of the monocular and binocular fusion based remote augmented reality follow-up perception system according to any one of claims 1-5, wherein: the method comprises the following steps:
step 1, placing an edge end single-binocular fusion follow-up intelligent sensing module in a cab of engineering machinery, and after ensuring that glass in front of the cab is not shielded, turning on a power supply of an edge end processor to enable the edge end processor to be in a waiting state, and waiting for establishing communication connection with a server end intelligent processing module;
step 2, turning on a power supply of the server side to enable the server side to be in a monitoring state, and waiting for establishing communication connection with the edge end single-binocular fusion follow-up intelligent sensing module and the user side augmented reality module;
step 3, an operator enters the console, wears the video glasses and starts remote control operation after the three-dimensional display equipment has a construction interface;
step 4, reading the head posture data of an operator by the edge end processor, controlling the servo cradle head to update the posture in real time, and simultaneously sending RGB information and depth data of a construction scene to the server end through a wireless transmission technology;
step 5, the server-side intelligent processing module carries out attitude estimation and bucket tip positioning of the bucket and environment map information construction through the received RGB information and depth data, and finally sends a fused image fused with the bucket attitude information, bucket tip position information and actual distance information between the bucket tip and objects around the dump truck to a user side for display;
step 6, the user side operator remotely controls the engineering machinery operation in a manner of being matched with the console through the fused image and the on-site three-dimensional information displayed by the video glasses, and meanwhile, the video glasses capture the head posture of the operator in real time and send the head posture to the server side to wait for the edge side processor to read;
and 7, repeating the step 4 to the step 6.
7. A bucket attitude tracking method based on sparse areas, wherein a monocular and binocular fusion based remote augmented reality follow-up perception system according to any one of claims 1 to 5 is used, characterized in that: the method comprises the following steps:
s1, placing a bucket under natural illumination, avoiding reflective objects around the bucket, taking 30 pictures by using photographing equipment to surround the bucket for a circle, and enabling the bucket to be located at the center of an image during photographing;
s2, opening RealityCapture software, and generating a bucket three-dimensional model by using 30 bucket photos, wherein the three-dimensional model is completely the same as a real bucket in proportion;
s3, placing virtual cameras at 2562 different positions around the three-dimensional model to render the three-dimensional model, acquiring sparse contour points of the bucket in the current posture by using a rendering map, back-projecting the contour points to a coordinate system of the bucket three-dimensional model for storage, and simultaneously storing normal vectors of the contour points and direction vectors of the current posture; finally, 2562 template views are generated;
s4, giving an initial posture of the bucket, multiplying the direction vectors of all the template views by the initial posture, and finding out the template view consistent with the initial posture; projecting the contour point of the template view onto a current real image, wherein the front 18 pixels of the contour point along the normal direction are designated as bucket pixels, the rear 18 pixels are designated as background pixels, and the bucket and the background are segmented to obtain the real contour of the bucket;
and S5, estimating the real posture of the bucket by using the distance between the model contour point and the real contour point along the normal direction, and further realizing the bucket tracking.
CN202211037134.1A 2022-08-26 2022-08-26 Remote augmented reality follow-up sensing system and method based on monocular and binocular fusion Active CN115514885B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211037134.1A CN115514885B (en) 2022-08-26 2022-08-26 Remote augmented reality follow-up sensing system and method based on monocular and binocular fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211037134.1A CN115514885B (en) 2022-08-26 2022-08-26 Remote augmented reality follow-up sensing system and method based on monocular and binocular fusion

Publications (2)

Publication Number Publication Date
CN115514885A true CN115514885A (en) 2022-12-23
CN115514885B CN115514885B (en) 2024-03-01

Family

ID=84501858

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211037134.1A Active CN115514885B (en) 2022-08-26 2022-08-26 Remote augmented reality follow-up sensing system and method based on monocular and binocular fusion

Country Status (1)

Country Link
CN (1) CN115514885B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117197769A (en) * 2023-11-03 2023-12-08 江苏智能无人装备产业创新中心有限公司 Loader front image generation system and method based on bucket position observation

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015199470A1 (en) * 2014-06-25 2015-12-30 한국과학기술원 Apparatus and method for estimating hand position utilizing head mounted color depth camera, and bare hand interaction system using same
CN107034942A (en) * 2017-05-24 2017-08-11 成都跟驰科技有限公司 Virtual reality system for excavator remote control
CN107888895A (en) * 2017-10-17 2018-04-06 三重机有限公司 Excavator tele-control system, method and excavator
CN107882103A (en) * 2017-10-26 2018-04-06 南京工业大学 Three-dimensional attitude display and remote automatic control system of excavator
US20180197345A1 (en) * 2016-09-13 2018-07-12 Youngzone Culture (Shanghai) Co., Ltd. Augmented reality technology-based handheld viewing device and method thereof
CN108797669A (en) * 2018-06-20 2018-11-13 清华大学 A kind of autonomous 3D excavations construction robot
CN109828658A (en) * 2018-12-17 2019-05-31 彭晓东 A kind of man-machine co-melting long-range situation intelligent perception system
US20200005520A1 (en) * 2018-06-27 2020-01-02 Shanghai United Imaging Healthcare Co., Ltd. Method and system for fusing image data
CN112116631A (en) * 2020-09-07 2020-12-22 江苏瑞科科技有限公司 Industrial augmented reality combined positioning system
CN112554253A (en) * 2020-11-27 2021-03-26 徐工集团工程机械有限公司 Multifunctional emergency rescue vehicle and control method thereof
WO2021098441A1 (en) * 2019-11-20 2021-05-27 Oppo广东移动通信有限公司 Hand posture estimation method and apparatus, device and computer storage medium
US20210311320A1 (en) * 2020-04-06 2021-10-07 Pike Enterprises, Llc Virtual reality tracking system
CN113723279A (en) * 2021-08-30 2021-11-30 东南大学 Multi-target tracking acceleration method based on time-space optimization in edge computing environment

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015199470A1 (en) * 2014-06-25 2015-12-30 한국과학기술원 Apparatus and method for estimating hand position utilizing head mounted color depth camera, and bare hand interaction system using same
US20180197345A1 (en) * 2016-09-13 2018-07-12 Youngzone Culture (Shanghai) Co., Ltd. Augmented reality technology-based handheld viewing device and method thereof
CN107034942A (en) * 2017-05-24 2017-08-11 成都跟驰科技有限公司 Virtual reality system for excavator remote control
CN107888895A (en) * 2017-10-17 2018-04-06 三重机有限公司 Excavator tele-control system, method and excavator
CN107882103A (en) * 2017-10-26 2018-04-06 南京工业大学 Three-dimensional attitude display and remote automatic control system of excavator
CN108797669A (en) * 2018-06-20 2018-11-13 清华大学 A kind of autonomous 3D excavations construction robot
US20200005520A1 (en) * 2018-06-27 2020-01-02 Shanghai United Imaging Healthcare Co., Ltd. Method and system for fusing image data
CN109828658A (en) * 2018-12-17 2019-05-31 彭晓东 A kind of man-machine co-melting long-range situation intelligent perception system
WO2021098441A1 (en) * 2019-11-20 2021-05-27 Oppo广东移动通信有限公司 Hand posture estimation method and apparatus, device and computer storage medium
US20210311320A1 (en) * 2020-04-06 2021-10-07 Pike Enterprises, Llc Virtual reality tracking system
CN112116631A (en) * 2020-09-07 2020-12-22 江苏瑞科科技有限公司 Industrial augmented reality combined positioning system
CN112554253A (en) * 2020-11-27 2021-03-26 徐工集团工程机械有限公司 Multifunctional emergency rescue vehicle and control method thereof
CN113723279A (en) * 2021-08-30 2021-11-30 东南大学 Multi-target tracking acceleration method based on time-space optimization in edge computing environment

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
陈熙霖等: "面向真实世界的智能感知与交互", 中国科学:信息科学, 31 August 2016 (2016-08-31) *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117197769A (en) * 2023-11-03 2023-12-08 江苏智能无人装备产业创新中心有限公司 Loader front image generation system and method based on bucket position observation
CN117197769B (en) * 2023-11-03 2024-01-26 江苏智能无人装备产业创新中心有限公司 Loader front image generation system and method based on bucket position observation

Also Published As

Publication number Publication date
CN115514885B (en) 2024-03-01

Similar Documents

Publication Publication Date Title
CN111062873B (en) Parallax image splicing and visualization method based on multiple pairs of binocular cameras
CN108780586B (en) Image processing method, display device and inspection system
CN110458897B (en) Multi-camera automatic calibration method and system and monitoring method and system
WO2019242262A1 (en) Augmented reality-based remote guidance method and device, terminal, and storage medium
TWI617277B (en) Pose estimation apparatus and vacuum cleaner system
CN106898022A (en) A kind of hand-held quick three-dimensional scanning system and method
US20170178392A1 (en) 3d scanning apparatus including scanning sensor detachable from screen
JP5093053B2 (en) Electronic camera
CN206162398U (en) A stereovision follow -up display system for heavy machinery is long -range unattended
CN112634318B (en) Teleoperation system and method for underwater maintenance robot
CN108093244B (en) Remote follow-up stereoscopic vision system
CN111383348A (en) Method for remotely and synchronously controlling robot through virtual reality
CN114608561A (en) Positioning and mapping method and system based on multi-sensor fusion
CN115514885B (en) Remote augmented reality follow-up sensing system and method based on monocular and binocular fusion
JP2015118442A (en) Information processor, information processing method, and program
CN113701750A (en) Fusion positioning system of underground multi-sensor
CN111947650A (en) Fusion positioning system and method based on optical tracking and inertial tracking
JP2016541042A (en) Method and system for providing position or motion information for controlling at least one function of a vehicle
WO2017155005A1 (en) Image processing method, display device, and inspection system
Stricker et al. From interactive to adaptive augmented reality
McMurrough et al. 3D point of gaze estimation using head-mounted RGB-D cameras
CN115890693A (en) Mobile explosion-removing device and method based on immersive perception and bionic control
Yang et al. Seeing as it happens: Real time 3D video event visualization
US20200005527A1 (en) Method and apparatus for constructing lighting environment representations of 3d scenes
JP2001239983A (en) Outfit assisting system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant