CN112649813B - Method for indoor safety inspection of important place, inspection equipment, robot and terminal - Google Patents
Method for indoor safety inspection of important place, inspection equipment, robot and terminal Download PDFInfo
- Publication number
- CN112649813B CN112649813B CN202011477726.6A CN202011477726A CN112649813B CN 112649813 B CN112649813 B CN 112649813B CN 202011477726 A CN202011477726 A CN 202011477726A CN 112649813 B CN112649813 B CN 112649813B
- Authority
- CN
- China
- Prior art keywords
- indoor
- plane
- fitting
- point
- dimensional
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/89—Lidar systems specially adapted for specific applications for mapping or imaging
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/006—Theoretical aspects
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
- G01S17/42—Simultaneous measurement of distance and other co-ordinates
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/86—Combinations of lidar systems with systems other than lidar, radar or sonar, e.g. with direction finders
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/4802—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Electromagnetism (AREA)
- Computer Networks & Wireless Communication (AREA)
- General Physics & Mathematics (AREA)
- Image Processing (AREA)
- Optical Radar Systems And Details Thereof (AREA)
Abstract
The invention belongs to the technical field of indoor and outdoor positioning, and discloses a method, inspection equipment, a robot and a terminal for indoor safety inspection of an important place, which are used for constructing data in a safety environment; indoor modeling of laser point cloud; acquiring coordinate information, field angle and azimuth angle information of patrol personnel or robots; calculating the area which can be scanned or observed by the sensor, and performing three-dimensional reconstruction on the collected laser point cloud data to obtain a real-time indoor three-dimensional profile; according to the real-time three-dimensional reconstruction of the acquisition area of the equipment, comparing the real-time three-dimensional reconstruction with model data in a safe state, and calculating the indoor space change condition between the two models through an algorithm; and marking the indoor change conditions of the key places by colors. The invention can accurately monitor the indoor environment of the key site, enhance the safety guarantee of the key site and better maintain the public safety and social stability. The method for indoor safety inspection of important places can accurately monitor the indoor environment of the important places, so that the change condition occurring in the important places can be found.
Description
Technical Field
The invention belongs to the technical field of indoor and outdoor positioning, and particularly relates to a method, inspection equipment, a robot and a terminal for indoor safety inspection of important places.
Background
In recent years, the invasion of illegal criminal behaviors causes serious damage to the security and stability of society and the security of lives and properties of people. The public security organ is an important functional department for handling emergencies and maintaining social stability, and plays an important role in security work. The video monitoring cooperation inspection personnel is one of the main means of the security protection work of the current public security department, the main early warning of the traditional video monitoring system is realized by the personnel through seeing the large screen of the monitoring display, a large amount of manpower investment is needed, meanwhile, the requirement on the concentration degree is very high, the effect of taking evidence afterwards can only be played under most conditions, and the efficiency of taking evidence afterwards is extremely low.
The security work of important places needs to realize the full-coverage inspection of the jurisdiction area, and needs to discover potential safety hazards as early as possible and deal with emergencies in time. The patrol area is large, and the environment is complex. The current working mechanism that adopts is that manpower inspection and video monitoring combine, relies on the policeman on duty naked eye discovery entirely, often has the fish of missing the net. Therefore, a new method for indoor security inspection in important places is needed.
Through the above analysis, the problems and defects of the prior art are as follows:
(1) the main early warning of traditional video monitoring system is realized through seeing the big screen of control display by the personnel, needs a large amount of human input, and to personnel's responsibility heart, concentration degree requirement is very high simultaneously, can only play the effect of collecting evidence afterwards under most of the circumstances, and the efficiency of collecting evidence afterwards is extremely low moreover.
(2) The current working mechanism that adopts is that manpower inspection and video monitoring combine, relies on the policeman on duty naked eye discovery entirely, often has the fish of missing the net.
(3) In the traditional security monitoring system, because the position is fixed, dead angles of a camera exist, the effective visual distance is narrow, and the system can realize autonomous movement, omnibearing, dead angle-free real-time monitoring. In addition, the invention is not influenced by the illumination effect.
The difficulty in solving the above problems and defects is:
(1) the indoor area environment is extremely complex and comprises dynamic objects and static objects, the traditional indoor laser point cloud modeling usually needs a whole-house scanning mode and a post-processing mode, and cannot achieve a real-time effect.
(2) Due to the particularity and the temporality of the guard area, video monitoring is not allowed or installed in some rooms in any condition, and only manual inspection can be carried out. The biggest problem of the manpower inspection is influenced by subjective awareness, experience, responsibility and external environment of inspection personnel.
(3) The invention also combines a three-dimensional geographic information system, can lead patrol personnel to directly find the problem when patrolling through the three-dimensional GIS visualization technology, and can lead the security center manager to see the inspection condition of each patrol personnel in real time.
The significance of solving the problems and the defects is as follows:
(1) and the inspection timeliness is improved.
(2) The problem that patrol personnel cannot find the problem due to subjective judgment and influence of external factors is prevented.
Disclosure of Invention
Aiming at the problems in the prior art, the invention provides a method, a device, a robot and a terminal for indoor security inspection in important places, aiming at solving the problem that the prior art needs to be recognized by patrol personnel through naked eyes in security work in key places.
The invention is realized in such a way that a method for the indoor safety inspection of the important place comprises the following steps:
step one, constructing data under a safe environment: under the premise of safety of indoor environment of important places, data acquisition is carried out on a guard area through mobile acquisition equipment or a mobile robot carrying a laser radar function, and acquired data are transmitted to a background server.
The laser radar mainly uses a laser as a transmitting light source, adopts active remote sensing equipment of a photoelectric detection technical means, and consists of a transmitting system, a receiving system, information processing and the like. It uses a laser to densely sample the indoor surface environment to produce high precision x, y, z measurements. A manageable, displayable, analyzable and shared discrete multipoint cloud data set can be generated.
The movable acquisition equipment or the movable robot with the laser radar function is used for carrying out movable acquisition in a guard area, and 360-degree all-dimensional scanning on the indoor space is required to be carried out as far as possible in order to ensure the integrity of the indoor three-dimensional surface space profile. The laser radar collects laser point clouds, and the collected laser point clouds can be processed into highly accurate geographic registration x, y and z coordinates through a Global Positioning System (GPS)/Inertial Navigation System (INS).
Step two, performing laser point cloud indoor modeling: an indoor three-dimensional vector outline is constructed through laser point cloud, and indoor vector outline data with coordinate information are directly loaded into a three-dimensional geographic information system.
And step three, when patrolling is needed, patrolling the security area through the mobile acquisition equipment or the mobile robot carrying the laser radar function. Coordinate information of patrol personnel or the robot is obtained through the positioning equipment, and meanwhile the field angle and the azimuth angle information of the sensor are obtained.
The mobile acquisition or robot can perform indoor positioning through WIFI or Bluetooth technology.
And fourthly, calculating the area which can be scanned or observed by the sensor in real time according to the position of the patrol personnel or the robot and the information of the field angle, the azimuth angle and the height of the sensor, and performing three-dimensional reconstruction on the collected laser point cloud data to obtain a real-time indoor three-dimensional profile.
And fifthly, carrying out real-time three-dimensional reconstruction according to the acquisition area of the equipment, comparing the real-time three-dimensional reconstruction with model data in a safe state, finding a change condition, and calculating the difference between the two models, namely the change condition of the indoor space, through VTK _ DIFFERENCE in a vtK vtkBooleanoperationPolyDataFilter algorithm.
And setting the indoor space profile to be A in the safety state and B in the patrol state. Separately establishing A, B BSP trees, then using the BSP tree of A to perform space division on B, and dividing the B into an inner part and an outer part; then, the BSP tree of the B is used for carrying out space segmentation on the A and dividing the A into an inner part and an outer part; finally, according to different Boolean operations, the invention adopts a difference solving mode to carry out parallel operation on the divided partial models according to different Boolean operations, and the change result of B relative to A can be obtained.
And step six, marking the indoor change conditions of the key places by colors, simultaneously loading the marked indoor change conditions into a three-dimensional geographic information system, and directly checking the indoor changed positions by handheld equipment.
Because the indoor space contour is provided with an actual coordinate system, the indoor space contour can be directly loaded into the three-dimensional geographic information system. And simultaneously, independently dyeing and loading the interpolation according to the two contour interpolation values obtained in the fifth step. According to the principle of the fourth step, the situation in the visible range can be loaded in real time.
Further, in the second step, the method for performing indoor modeling of laser point cloud includes:
(1) and continuously generating a surface model from the discrete point cloud data to simulate the surface of an indoor scene, fitting each point set originally belonging to the same plane to determine a central plane, and simultaneously determining the outline of the plane. The three-dimensional coordinate information of the special line and point is extracted by utilizing the mutual relation of a plurality of planes with clear edge contour information and space position information.
Fitting a plane equation expression: ax + By + Cz + D ═ 0(C ≠ 0);
Then z is a0x+a1y+a2;
Knowing a set of n points (X) of datai,Yi,Zi),i=0,1,2,3,……n-1;
namely:
a0∑xi+a1∑yi+a2n=∑zi;
the combination and arrangement are as follows:
solving the linear equation to obtain a0,a1,a2。
(2) Determining the outline of the indoor space object on the plane:
the projected points from each point of the original real object to the fitting plane are obtained by solving the vertical line and the foothold of the point and the plane, and the contour line is fitted through all the projected points.
Suppose point A is any point in the material point set M and the coordinate value is (X)A,YA,ZA) The point a is the foot of the point A to the fitting plane, and the coordinate value is (X)a,Ya,Za) The known planar equation is z ═ a0x+a1y+a2The normal vector of which is (a)0,a1,-1)。
Then the equation for the perpendicular Aa is:
the coordinates of any point on Aa are as follows: (a)0K+XA,a1K+YA,-K+ZA);
Substituting the coordinate value into the fitting plane equation can solve:
the coordinate value of the point a is:
the coordinates of each point in the set N of drop foot points from all points to the fitting plane are obtained in the same way.
(3) Boundary fitting
Since the indoor environment is too complex, the curve boundary is fitted by a polynomial method, and the plane equation z is known to be a0x+a1y+a2The polynomial being y ═ w0+w1x+w2x2+…+wnxn. Wherein w0,w1,w2,……,wnAre coefficients of a polynomial.
X, Y coordinates for a set of curve boundary points:
……
in the least squares sense:
minimum w0,w1,w2,……,wnAnd the matrix form of the determined polynomial obtained through mathematical processing is as follows:
then using the boundary curve belonging to the plane z ═ a0x+a1y+a2The coefficient w can be uniquely determined0,w1,...,wnI.e. determine this curve.
Further, in the fourth step, the area which can be scanned or observed by the sensor calculated in real time is the acquisition area of the laser point cloud data.
Another object of the present invention is to provide an indoor security inspection apparatus for a key site that implements the method for indoor security inspection for a key site.
It is another object of the present invention to provide a computer-readable storage medium storing instructions that, when executed on a computer, cause the computer to perform the method for indoor security inspection of a key location.
It is a further object of the invention to provide a computer arrangement comprising a memory and a processor, the memory storing a computer program which, when executed by the processor, causes the processor to perform the method of indoor security check of a critical location.
Another object of the present invention is to provide a lidar-mounted mobile robot that operates the method for indoor security inspection of important places for indoor security inspection patrol.
Another object of the present invention is to provide a security check information data processing terminal for implementing the method for indoor security check of important places, which is a panoramic picture generation method.
By combining all the technical schemes, the invention has the advantages and positive effects that: the method for indoor safety inspection of important places can accurately monitor the indoor environment of the important places, so that the change condition occurring in the important places can be found. The invention is applied to key security places such as government buildings and the like, can further enhance the security guarantee for the key places, better maintains public security and social stability, promotes the security guarantee work of a public security organ in the key places to be changed from a passive corresponding type to an active guaranteeing type, and is an important step of intelligent emergency command of the public security organ.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the embodiments of the present invention will be briefly described below, and it is obvious that the drawings described below are only some embodiments of the present invention, and it is obvious for those skilled in the art that other drawings can be obtained according to the drawings without creative efforts.
Fig. 1 is a flowchart of a method for indoor security inspection of an important place according to an embodiment of the present invention.
Fig. 2 is a schematic view of a visible area provided by an embodiment of the present invention.
Fig. 3 is a camera orientation diagram provided by an embodiment of the invention.
Fig. 4 is a schematic point cloud diagram of an indoor object in a safe state according to an embodiment of the present invention.
Fig. 5 is a schematic diagram of the object contour after calculation and fitting in the safe state according to the embodiment of the present invention.
Fig. 6 is a schematic point cloud during the second scanning according to the embodiment of the present invention.
FIG. 7 is a schematic diagram of the object profile of the second fitting provided by the embodiment of the invention.
FIG. 8 is a graph showing the results of 2 comparisons before and after the example of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described in detail with reference to the following embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
In view of the problems in the prior art, the present invention provides a method, an inspection apparatus, a robot and a terminal for indoor security inspection in an important place, and the present invention is described in detail below with reference to the accompanying drawings.
As shown in fig. 1, the method for indoor security inspection of an important place provided by the embodiment of the present invention includes the following steps:
s101, constructing data under a secure environment: under the premise of safety of indoor environment of important places, data acquisition is carried out on a guard area through mobile acquisition equipment or a mobile robot carrying a laser radar function, and acquired data are transmitted to a background server.
S102, constructing an indoor three-dimensional vector outline through laser point cloud, and directly loading indoor vector outline data with coordinate information into a three-dimensional geographic information system.
S103, when patrolling is needed, patrolling the security area through the mobile acquisition equipment or the mobile robot carrying the laser radar function. Coordinate information of patrol personnel or the robot is obtained through the positioning equipment, and meanwhile the field angle and the azimuth angle information of the sensor are obtained.
And S104, calculating the area which can be scanned or observed by the sensor in real time, namely the acquisition area of the laser point cloud data according to the position of the patrol personnel or the robot and the information of the field angle, the azimuth angle and the height of the sensor. Meanwhile, three-dimensional reconstruction is carried out on the collected laser point cloud data to obtain a real-time indoor three-dimensional profile.
Fig. 2 is a schematic view of a visible area according to an embodiment of the present invention, which is provided to simulate a current visible area of a sensor in a three-dimensional geographic information system by constructing a polygonal line and a polygon. Fig. 3 is a camera orientation diagram of the present invention.
In this embodiment, the angle of view is as shown in fig. 2, and the left side is the position of the sensor, so that the current visible area of the sensor can be simulated in the three-dimensional geographic information system by constructing a polygonal line and a polygon according to the data of the angle of view.
In this embodiment, the azimuth angle is shown in fig. 3, Yaw is the Yaw angle in the horizontal direction, Pitch is the Pitch angle in the vertical direction, Roll is the rotation angle, and the spatial position and attitude of the sensor can be accurately described by matching the X, Y, and Z coordinates of the sensor. Meanwhile, the visual area of the current sensor can be accurately visualized and simulated in the three-dimensional geographic information system by combining the field angle of the sensor.
S105, performing real-time three-dimensional reconstruction according to the acquisition area of the equipment, comparing the real-time three-dimensional reconstruction with model data in a safe state, finding a change condition, and calculating the difference between two models, namely the condition of indoor space change, through VTK _ DIFFERENCE in a vtK vtkBooleanoperationPolyDataFilter algorithm.
And S106, marking the indoor change conditions of the key places by colors, simultaneously loading the marked indoor change conditions into the three-dimensional geographic information system, and directly checking the indoor changed positions by the handheld equipment.
The present invention will be further described with reference to the following examples.
The method for indoor safety inspection of important places provided by the embodiment of the invention comprises the following steps:
the first step is as follows: and constructing data in a safe environment.
Under the premise of safety of indoor environment in important places, data acquisition is carried out on a guard area through a patrol worker holding mobile acquisition equipment or a mobile robot carrying a laser radar function, and acquired data are transmitted to a background server.
The second step is that: and carrying out indoor modeling of laser point cloud.
(1) The point cloud data of the indoor laser is discrete, cannot accurately represent geometric information of a real scene, and cannot meet practical application, so that the discrete point cloud data needs to be continuously generated into a surface model to simulate the surface of the indoor scene. However, due to the error caused by laser scanning, the points that should be located on the same plane will show a random distribution with a slight deviation from the center of a certain plane.
In practical use, each point set originally belonging to the same plane needs to be fitted to determine the central plane and determine the outline of the plane at the same time. The three-dimensional coordinate information of the special line and point is extracted by utilizing the mutual relation of a plurality of planes with clear edge contour information and space position information.
Fitting a plane equation expression: ax + By + Cz + D ═ 0(C ≠ 0);
Then z is a0x+a1y+a2;
Knowing a set of n points (X) of datai,Yi,Zi),i=1,2,3,……n-1;
namely:
comprises the following steps:
a0∑xi+a1∑yi+a2n=∑zi;
solving the linear equation to obtain a0,a1,a2。
(2) Only the equation of the plane is obtained at this time, and the size of the actual object cannot be determined, which requires determining the outline of the indoor space real object on the plane.
Obtaining the projection profile of the real object on the plane first requires finding the projection point from each point of the original real object to the fitting plane, which can be obtained by finding the perpendicular line and the foot between the point and the plane. Then the contour line is fitted through all the projection points.
Suppose point A is any point in the material point set M and the coordinate value is (X)A,YA,ZA) The point a is the foot of the point A to the fitting plane, and the coordinate value is (X)a,Ya,Za) The known planar equation is z ═ a0x+a1y+a2The normal vector of which is (a)0,a1,-1)。
Then the equation for the perpendicular Aa is:
the coordinates of any point on Aa are as follows: (a)0K+XA,a1K+YA,-K+ZA);
Substituting the coordinate value into the fitting plane equation can solve:
the coordinate value of the point a is:
then the coordinates of all points to each point in the set N of drop foot points of the fitted plane can be obtained in the same way.
(3) Boundary fitting
Since the indoor environment is too complex, the curve boundary is fitted by a polynomial method, and the plane equation z is known to be a0x+a1y+a2The polynomial being y ═ w0+w1x+w2x2+…+wnxn。
X, Y coordinates for a set of curve boundary points:
……
in the least squares sense:
minimum w0,w1,w2,……,wnAnd the matrix form of the determined polynomial obtained through mathematical processing is as follows:
then using the boundary curve belonging to the plane z ═ a0x+a1y+a2The coefficient w can be uniquely determined0,w1,...,wnI.e. determine this curve.
(4) Therefore, an indoor three-dimensional vector outline can be constructed through the laser point cloud.
The third step:
the indoor vector outline data obtained by the second step of processing has coordinate information, so that the indoor vector outline data can be directly loaded into a three-dimensional geographic information system.
The fourth step:
when patrolling is needed, patrolling is conducted on a security area through a patrol person holding the mobile acquisition equipment or a mobile robot carrying a laser radar function. Coordinate information of patrol personnel or the robot is obtained through the positioning equipment, and meanwhile the field angle and the azimuth angle information of the sensor can be obtained.
The fifth step:
according to the position of the patrol personnel or the robot, the area which can be scanned or observed by the sensor can be calculated in real time by combining the information of the angle of view, the azimuth angle, the height and the like of the sensor. Meanwhile, the laser point cloud data acquisition area is also provided. And performing three-dimensional reconstruction on the laser point cloud data acquired at this time according to the content of the second step to obtain a real-time indoor three-dimensional profile.
And a sixth step:
and (3) according to the real-time three-dimensional reconstruction of the acquisition area of the equipment, comparing the real-time three-dimensional reconstruction with model data in a safe state, finding a change condition, and calculating the difference between the two models, namely the change condition of the indoor space, through VTK _ DIFFERENCE in a vtK vtkBooleanoperationPolyDataFilter algorithm.
The seventh step:
the indoor change conditions of key places are marked through colors and loaded into the three-dimensional geographic information system, and patrol personnel can directly check the indoor changed positions through handheld equipment, so that the patrol personnel can conveniently and immediately check the indoor changed positions.
Fig. 4 is a point cloud of indoor objects in a safe state; FIG. 5 illustrates a calculated fit of the object profile in the safe state; FIG. 6 point cloud at the second scan; FIG. 7 is a second fit of the object profile; in fig. 7, the object on the table top should be located at the left side of the drawing under the normal condition, but after the second scanning comparison, the actual object is changed in position.
It should be noted that the embodiments of the present invention can be realized by hardware, software, or a combination of software and hardware. The hardware portion may be implemented using dedicated logic; the software portions may be stored in a memory and executed by a suitable instruction execution system, such as a microprocessor or specially designed hardware. Those skilled in the art will appreciate that the apparatus and methods described above may be implemented using computer executable instructions and/or embodied in processor control code, such code being provided on a carrier medium such as a disk, CD-or DVD-ROM, programmable memory such as read only memory (firmware), or a data carrier such as an optical or electronic signal carrier, for example. The apparatus and its modules of the present invention may be implemented by hardware circuits such as very large scale integrated circuits or gate arrays, semiconductors such as logic chips, transistors, or programmable hardware devices such as field programmable gate arrays, programmable logic devices, etc., or by software executed by various types of processors, or by a combination of hardware circuits and software, e.g., firmware.
The above description is only for the purpose of illustrating the present invention and the appended claims are not to be construed as limiting the scope of the invention, which is intended to cover all modifications, equivalents and improvements that are within the spirit and scope of the invention as defined by the appended claims.
Claims (9)
1. A method for indoor security inspection of a key location, the method comprising:
carrying out data acquisition on an indoor security area through mobile acquisition equipment or a mobile robot carrying a laser radar, and transmitting the acquired data to a background server;
constructing an indoor three-dimensional vector outline through laser point cloud, and directly loading indoor vector outline data with coordinate information into a three-dimensional geographic information system;
coordinate information of patrol personnel or patrol mobile robots is obtained through positioning equipment, and meanwhile field angle and azimuth angle information of the sensor is obtained;
according to the position of patrol personnel or a patrol mobile robot, the area scanned or detected by the sensor is calculated in real time by combining the information of the field angle, the azimuth angle and the height of the sensor, and the collected laser point cloud data is subjected to three-dimensional reconstruction to obtain a real-time indoor three-dimensional profile;
comparing the real-time indoor three-dimensional contour with model data in a safe state according to the obtained real-time indoor three-dimensional contour, and calculating the difference between the real-time indoor three-dimensional contour and the model in the safe state to obtain the indoor space change condition;
marking the obtained indoor change conditions of the key places by colors, loading the indoor change conditions into a three-dimensional geographic information system, and checking the indoor changed positions by handheld equipment;
the method is characterized in that an indoor three-dimensional vector outline is constructed through laser point cloud, and indoor vector outline data with coordinate information are directly loaded into a three-dimensional geographic information system, and the method comprises the following steps:
(1) continuously generating a surface model from the discrete point cloud data to simulate the surface of an indoor scene, fitting each point set originally belonging to the same plane to determine a central plane, and simultaneously determining the outline of the plane; extracting special three-dimensional coordinate information of lines and points by utilizing the interrelation of a plurality of planes with definite edge contour information and space position information;
(2) determining the contour of the indoor space real object on the plane, obtaining projection points from each point of the original real object to a fitting plane by solving the vertical line and the foothold of the point and the plane, and fitting the contour line of the real object through all the projection points;
(3) and (4) boundary fitting, wherein a polynomial method is selected for fitting at the boundary of the curve due to the fact that the indoor environment is too complex.
2. The method for indoor security inspection of a significant site according to claim 1, wherein the fitting method is:
(1) continuously generating a surface model from the discrete point cloud data to simulate the surface of an indoor scene, fitting each point set originally belonging to the same plane to determine a central plane, and simultaneously determining the outline of the plane; extracting special three-dimensional coordinate information of lines and points by utilizing the interrelation of a plurality of planes with definite edge contour information and space position information;
the specific method comprises the following steps:
fitting a plane equation expression: ax + By + Cz + D is 0, C is not equal to 0;
Then z is a0x+a1y+a2;
Knowing a set of n points (X) of datai,Yi,Zi),i=0,1,2,3,……n-1;
namely:
a0∑xi+a1∑yi+a2n=∑zi;
the combination and arrangement are as follows:
solving the linear equation to obtain a0,a1,a2The plane can be confirmed;
(2) determining the contour of the indoor space real object on the plane, obtaining projection points from each point of the original real object to a fitting plane by solving the vertical line and the foothold of the point and the plane, and fitting the contour line of the real object through all the projection points;
the specific method comprises the following steps:
suppose point A is any point in the material point set M and the coordinate value is (X)A,YA,ZA) The point a is the foot of the point A to the fitting plane, and the coordinate value is (X)a,Ya,Za) The known planar equation is z ═ a0x+a1y+a2The normal vector of which is (a)0,a1,-1),
Then the equation for the perpendicular Aa is:
the coordinates of any point on Aa are as follows: (a)0K+XA,a1K+YA,-K+ZA);
Substituting the coordinate value into the fitting plane equation can solve:
the coordinate value of the point a is:
obtaining the coordinates of each point in the set N of the foot points from all the points to the fitting plane by the same method;
(3) boundary fitting, namely fitting by a polynomial method for curve boundary selection because the indoor environment is too complex;
the specific method comprises the following steps:
known plane equation z ═ a0x+a1y+a2The polynomial being y ═ w0+w1x+w2x2+…+wnxnWherein w is0,w1,w2,……,wnIs a coefficient of a polynomial;
x, Y coordinates for a set of curve boundary points:
in the least squares sense:
minimum w0,w1,w2,……,wnAnd the matrix form of the determined polynomial obtained through mathematical processing is as follows:
then using the curve boundary belonging to the plane z ═ a0x+a1y+a2Then can be uniquely determinedCoefficient w0,w1,...,wnI.e. determine this curve.
3. The method for indoor security inspection of a public place of interest as set forth in claim 1, wherein the area scanned or detected by the sensor calculated in real time is an acquisition area of laser point cloud data.
4. An indoor security inspection apparatus for a public space, which performs the method for indoor security inspection for a public space according to any one of claims 1 to 3.
5. A computer-readable storage medium storing instructions that, when executed on a computer, cause the computer to perform the method of security inspection of a venue indoor as claimed in any one of claims 1 to 3.
6. A computer arrangement, characterized in that the computer arrangement comprises a memory and a processor, the memory storing a computer program which, when executed by the processor, causes the processor to carry out the method of indoor security check of a significant venue as claimed in any one of claims 1 to 3.
7. A mobile robot carrying a laser radar, wherein the mobile robot carrying a laser radar operates the method for indoor security inspection of a critical place according to any one of claims 1 to 3, and is used for patrol of indoor security inspection.
8. A security inspection information data processing terminal, characterized in that the security inspection information data processing terminal is used for realizing the panoramic picture generation method of the indoor security inspection method of the important place of any claim 1 to 3.
9. An intelligent home system is characterized in that an intelligent home system end is used for realizing the panoramic picture generation method for the indoor safety inspection of the important places according to any one of claims 1 to 3.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011477726.6A CN112649813B (en) | 2020-12-15 | 2020-12-15 | Method for indoor safety inspection of important place, inspection equipment, robot and terminal |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011477726.6A CN112649813B (en) | 2020-12-15 | 2020-12-15 | Method for indoor safety inspection of important place, inspection equipment, robot and terminal |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112649813A CN112649813A (en) | 2021-04-13 |
CN112649813B true CN112649813B (en) | 2022-02-08 |
Family
ID=75355424
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011477726.6A Active CN112649813B (en) | 2020-12-15 | 2020-12-15 | Method for indoor safety inspection of important place, inspection equipment, robot and terminal |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112649813B (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106495005A (en) * | 2016-12-23 | 2017-03-15 | 大连华锐重工起重机有限公司 | A kind of crane collision resistant detecting and controlling system |
CN108413965A (en) * | 2018-03-12 | 2018-08-17 | 中国矿业大学 | A kind of indoor and outdoor crusing robot integrated system and crusing robot air navigation aid |
CN111524103A (en) * | 2020-04-10 | 2020-08-11 | 山东科技大学 | Circular tunnel central axis extraction method based on three-dimensional laser point cloud |
CN111932671A (en) * | 2020-08-22 | 2020-11-13 | 扆亮海 | Three-dimensional solid model reconstruction method based on dense point cloud data |
Family Cites Families (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5652717A (en) * | 1994-08-04 | 1997-07-29 | City Of Scottsdale | Apparatus and method for collecting, analyzing and presenting geographical information |
WO2006027339A2 (en) * | 2004-09-06 | 2006-03-16 | The European Community, Represented By The European Commission | Method and system for 3d scene change detection |
CN102176252B (en) * | 2011-01-26 | 2013-02-13 | 新疆中钜电子科技有限公司 | Equipment and facility safety inspection intelligent patrolling operation system and application method thereof |
CN104658152B (en) * | 2015-02-15 | 2017-10-20 | 西安交通大学 | A kind of moving object intrusion alarm method based on video |
CN105844629B (en) * | 2016-03-21 | 2018-12-18 | 河南理工大学 | A kind of large scene City Building facade point cloud automatic division method |
CN106097474A (en) * | 2016-06-08 | 2016-11-09 | 朱兰英 | System is analyzed in a kind of indoor patrol based on real-time virtual reality technology |
CN106652291A (en) * | 2016-12-09 | 2017-05-10 | 华南理工大学 | Indoor simple monitoring and alarming system and method based on Kinect |
CN106953784B (en) * | 2017-03-17 | 2021-09-10 | 智科达(厦门)科技有限公司 | Indoor environment intrusion detection method in smart home |
CN107977992A (en) * | 2017-12-05 | 2018-05-01 | 深圳大学 | A kind of building change detecting method and device based on unmanned plane laser radar |
CN110018496A (en) * | 2018-01-10 | 2019-07-16 | 北京京东尚科信息技术有限公司 | Obstacle recognition method and device, electronic equipment, storage medium |
SG11202011825TA (en) * | 2018-05-30 | 2020-12-30 | Vi3D Labs Inc | Three-dimensional surface scanning |
CN109387858A (en) * | 2018-08-29 | 2019-02-26 | 北京信息科技大学 | A kind of full filed angle solid-state laser radar detection apparatus and obstacle detection method |
US10846511B2 (en) * | 2018-12-20 | 2020-11-24 | Here Global B.V. | Automatic detection and positioning of pole-like objects in 3D |
CN109949326B (en) * | 2019-03-21 | 2020-09-08 | 苏州工业园区测绘地理信息有限公司 | Building contour line extraction method based on knapsack type three-dimensional laser point cloud data |
CN110580459B (en) * | 2019-08-27 | 2022-02-18 | 浙江大华技术股份有限公司 | Safety check method and control equipment |
CN111402402B (en) * | 2020-03-14 | 2022-04-05 | 招商局重庆交通科研设计院有限公司 | Three-dimensional visual modeling method for highway side slope |
CN111784039B (en) * | 2020-06-28 | 2020-12-25 | 中国人民公安大学 | City large-scale activity police force defense deploying method based on dynamic risk assessment |
CN111928839A (en) * | 2020-08-07 | 2020-11-13 | 北京星天地信息科技有限公司 | Method and device for planning passing route and computer equipment |
-
2020
- 2020-12-15 CN CN202011477726.6A patent/CN112649813B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106495005A (en) * | 2016-12-23 | 2017-03-15 | 大连华锐重工起重机有限公司 | A kind of crane collision resistant detecting and controlling system |
CN108413965A (en) * | 2018-03-12 | 2018-08-17 | 中国矿业大学 | A kind of indoor and outdoor crusing robot integrated system and crusing robot air navigation aid |
CN111524103A (en) * | 2020-04-10 | 2020-08-11 | 山东科技大学 | Circular tunnel central axis extraction method based on three-dimensional laser point cloud |
CN111932671A (en) * | 2020-08-22 | 2020-11-13 | 扆亮海 | Three-dimensional solid model reconstruction method based on dense point cloud data |
Also Published As
Publication number | Publication date |
---|---|
CN112649813A (en) | 2021-04-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111459166B (en) | Scene map construction method containing trapped person position information in post-disaster rescue environment | |
US11842516B2 (en) | Homography through satellite image matching | |
CN112489099B (en) | Point cloud registration method and device, storage medium and electronic equipment | |
Kopsida et al. | Markerless BIM registration for mobile augmented reality based inspection | |
CN111192321A (en) | Three-dimensional positioning method and device for target object | |
CN112802208B (en) | Three-dimensional visualization method and device in terminal building | |
Chung et al. | Development of LiDAR simulator for backpack-mounted mobile indoor mapping system | |
Chen et al. | Real-time 3D mobile mapping for the built environment | |
CN116363585A (en) | On-line monitoring method and system for power transmission line | |
CN114066985B (en) | Method for calculating hidden danger distance of power transmission line and terminal | |
CN111444570A (en) | Construction error information acquisition method and device | |
CN114821395A (en) | Abnormity positioning method, device, equipment and readable storage medium | |
WO2022025283A1 (en) | Measurement processing device, method, and program | |
Gimeno et al. | An augmented reality (AR) CAD system at construction sites | |
CN112669392B (en) | Map positioning method and system applied to indoor video monitoring system | |
CN112184903B (en) | Method, device, equipment and medium for detecting high-voltage line tree obstacle risk points | |
CN112649813B (en) | Method for indoor safety inspection of important place, inspection equipment, robot and terminal | |
CN111156983B (en) | Target equipment positioning method, device, storage medium and computer equipment | |
Sitnik et al. | Monitoring surface degradation process by 3D structured light scanning | |
Tsoy et al. | Exhaustive simulation approach for a virtual camera calibration evaluation in gazebo | |
CN111597954A (en) | Method and system for identifying vehicle position in monitoring video | |
CN116203976A (en) | Indoor inspection method and device for transformer substation, unmanned aerial vehicle and storage medium | |
CN115982824A (en) | Construction site worker space management method and device, electronic equipment and storage medium | |
Ma et al. | Low‐Altitude Photogrammetry and Remote Sensing in UAV for Improving Mapping Accuracy | |
Wang et al. | Rapid in-flight image quality check for UAV-enabled bridge inspection |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |