WO2022219780A1 - Inspection assistance device, inspection assistance system, method for assisting inspection, and inspection assistance program - Google Patents

Inspection assistance device, inspection assistance system, method for assisting inspection, and inspection assistance program Download PDF

Info

Publication number
WO2022219780A1
WO2022219780A1 PCT/JP2021/015597 JP2021015597W WO2022219780A1 WO 2022219780 A1 WO2022219780 A1 WO 2022219780A1 JP 2021015597 W JP2021015597 W JP 2021015597W WO 2022219780 A1 WO2022219780 A1 WO 2022219780A1
Authority
WO
WIPO (PCT)
Prior art keywords
inspection
inspection target
orientation
camera
display
Prior art date
Application number
PCT/JP2021/015597
Other languages
French (fr)
Japanese (ja)
Inventor
隆博 加島
愛梨 守谷
健央 川浦
琴由 笹山
結佳 高橋
智也 上船
淳志 堀
Original Assignee
三菱電機株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 三菱電機株式会社 filed Critical 三菱電機株式会社
Priority to JP2023514277A priority Critical patent/JP7361992B2/en
Priority to PCT/JP2021/015597 priority patent/WO2022219780A1/en
Publication of WO2022219780A1 publication Critical patent/WO2022219780A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast

Definitions

  • the present disclosure relates to an inspection support device, an inspection support system, an inspection support method, and an inspection support program.
  • a mobile terminal to which this technology is applied detects an inspection target equipped with a reference marker, acquires registration information of the detected inspection target from the management database, and displays the acquired information superimposed on the camera image. can be done.
  • JP 2020-80147 (for example, paragraphs 0012 to 0037)
  • An object of the present disclosure is to provide an inspection support device, an inspection support system, an inspection support method, and an inspection support program that can support inspection work without requiring preparatory work.
  • An inspection support device based on at least one of a camera image captured by a camera provided in a mobile terminal and a detection signal obtained by an inertial sensor provided in the mobile terminal, determines three-dimensional coordinates a terminal position/orientation estimation unit that estimates the position and orientation of the mobile terminal in the system; an inspection object detection unit that detects an inspection object from a camera image captured by the camera; an inspection target tracking unit that tracks the indicated inspection target area in chronological order of photographing time and identifies the same inspection target area in a plurality of camera images arranged in time series; an inspection target position estimating unit for estimating the coordinates of the inspection target region in the three-dimensional coordinate system based on the inspection target region in at least two camera images out of the camera images; and a display control unit that causes a display device to display a display component to be used.
  • An inspection support method of the present disclosure is a method executed by an inspection support device, and includes a camera image captured by a camera provided in a mobile terminal and a detection signal obtained by an inertial sensor provided in the mobile terminal. estimating the position and orientation of the mobile terminal in a three-dimensional coordinate system based on at least one of; detecting an inspection target from a camera image captured by the camera; a step of tracking the inspection target area indicating the inspection target in chronological order of photographing time, and identifying the same inspection target area in a plurality of camera images arranged in time series; estimating the coordinates of the inspection target area in the three-dimensional coordinate system based on the inspection target area in at least two camera images out of the camera images; and making the inspection target area visible. and a step of displaying the display component on a display device.
  • inspection work can be supported without the need for preparatory work.
  • FIG. 1 is a functional block diagram schematically showing configurations of an inspection support device and a portable terminal according to Embodiment 1;
  • FIG. FIG. 2 is a diagram showing an example of a terminal device and an inspection target according to Embodiment 1;
  • FIG. 4 is a diagram showing a display example of the display device of the terminal device according to Embodiment 1;
  • FIG. 2 is a diagram illustrating an example of hardware configurations of an inspection support device and a mobile terminal according to Embodiment 1;
  • FIG. 3 is a diagram showing another example of the hardware configuration of the inspection support device and the mobile terminal according to Embodiment 1;
  • FIG. 7 is a flowchart showing a terminal position/orientation estimation thread by the inspection support device according to the first embodiment; 7 is a flow chart showing an inspection target acquisition thread by the inspection support device according to the first embodiment;
  • FIG. 10 is a functional block diagram schematically showing configurations of an inspection support device and a portable terminal according to Embodiment 2;
  • FIG. 10 is a diagram showing a display example of a camera image of the terminal device according to Embodiment 2 (at the time of reading a character string);
  • FIG. 11 is a diagram showing a display example of a display device of a terminal device according to Embodiment 2;
  • FIG. 11 is a functional block diagram schematically showing the configuration of an inspection support device and a portable terminal according to Embodiment 3;
  • FIG. 11 is a diagram showing a display example of a display device of a terminal device according to Embodiment 3;
  • FIG. 11 is a functional block diagram schematically showing the configuration of an inspection support device and a portable terminal according to Embodiment 4;
  • FIG. 11 is a functional block diagram schematically showing the configuration of an inspection support device and a portable terminal according to Embodiment 5;
  • FIG. 12 is a functional block diagram schematically showing the configuration of an inspection support device and a portable terminal according to Embodiment 6;
  • FIG. 13 is a diagram showing a display example of a display device of a terminal device according to Embodiment 6;
  • FIG. 14 is a functional block diagram schematically showing the configuration of an inspection support device and a portable terminal according to Embodiment 7;
  • FIG. 12 is a diagram showing a display example of a display device of a terminal device according to Embodiment 7;
  • FIG. 1 is a functional block diagram schematically showing configurations of an inspection support device 100 and a mobile terminal 10 according to Embodiment 1.
  • the inspection support device 100 is a device capable of executing the inspection support method according to the first embodiment.
  • the inspection support device 100 can execute an inspection support method by executing an inspection support program.
  • the mobile terminal 10 has an inspection support device 100, a camera 110, and a display device 180.
  • Camera 110 is, for example, an RGB camera.
  • Camera 110 may be a stereo camera.
  • the display device 180 is, for example, a display device such as a liquid crystal display device.
  • the display device 180 may be a touch panel that has a function as a display device and a function as an input unit that is a user operation unit.
  • the mobile terminal 10 may also include an inertial sensor (described later with reference to FIG. 4).
  • inspection support device 100 includes terminal position/orientation estimation unit 120 , inspection object detection unit 130 , inspection object tracking unit 140 , inspection object position estimation unit 150 , and display device 180 . and a display control unit 170 for displaying.
  • the inspection support device 100 may have an inspection target storage unit 160 that is a storage device that stores information.
  • the terminal position and orientation estimation unit 120 estimates the position and orientation (that is, the position and orientation) of the mobile terminal in a three-dimensional coordinate system representing the world space from the camera image captured by the camera 110 .
  • This three-dimensional coordinate system does not need to be a coordinate system whose origin is a predetermined position (for example, predetermined longitude and latitude on the earth or a predetermined building point).
  • the three-dimensional coordinate system does not have to be a predetermined coordinate system. It may be determined based on the position and orientation of the mobile terminal 10 .
  • the terminal position/orientation estimating unit 120 calculates the position and orientation of the mobile terminal based on the position and orientation of the mobile terminal when the mobile terminal is activated (that is, when the device is activated) or when the inspection support program is activated (that is, when the application is activated).
  • a dimensional coordinate system may be set. Therefore, by integrating the change in the position and orientation of the mobile terminal from the position and orientation of the mobile terminal at the time of startup (that is, at the time of device startup or application startup), a three-dimensional coordinate system with the position at the time of startup as the origin , the position and orientation of the mobile terminal at each time can be estimated.
  • an estimation method there is a method of calculating the position and orientation at each time based on the information of the feature amount in the camera image, and a method of calculating the position and orientation at each time based on information such as the velocity from the acceleration sensor and the rotation from the angular velocity sensor.
  • An existing method can be used as a method for estimating the position and orientation.
  • the terminal position/orientation estimation unit 120 detects at least one of a camera image captured by the camera 110 and a detection signal obtained by the inertial sensor. , the position and orientation of the mobile terminal 10 in the three-dimensional coordinate system may be estimated. In addition, terminal position/orientation estimation section 120 performs 3 A position and orientation of the mobile terminal (10) in a dimensional coordinate system may be estimated.
  • the inspection target detection unit 130 detects inspection targets (eg, valves, gauges, etc.) from the camera image captured by the camera 110 .
  • inspection targets eg, valves, gauges, etc.
  • the inspection target tracking unit 140 tracks the inspection target area indicating the inspection target detected in the multiple camera images arranged in time series, and identifies the same inspection target area in the multiple camera images arranged in time series.
  • the inspection target tracking unit 140 compares the inspection target region in the current camera image detected by the inspection target detection unit 130 with the inspection target region in the previous detection, thereby determining the inspection target region on the camera image. Track position changes over time. Since the inspection target detection unit 130 performs detection processing for each camera image (that is, each frame of the camera image), the relationship between the inspection target areas detected in the camera images before and after the time is unknown. Through the processing of the inspection object tracking unit 140, the relationship between the inspection object areas detected in the camera images before and after the time is grasped.
  • the inspection target tracking unit 140 determines that the same inspection target is indicated when a part of the detected inspection target region rectangle overlaps between the camera images before and after the time.
  • the inspection target tracking unit 140 uses a technique called optical flow to calculate the movement direction and amount of movement of pixels between camera images, and compares the previous and subsequent inspection target areas with the movement of pixels, thereby performing an inspection. Able to identify relationships between regions of interest.
  • the inspection target position estimation unit 150 calculates the coordinates of the inspection target region in the three-dimensional coordinate system based on the position and orientation of the mobile terminal 10 and the inspection target regions in at least two camera images out of the plurality of camera images. presume.
  • the inspection target storage unit 160 stores the coordinates of the inspection target area indicating the inspection target.
  • the display control unit 170 causes the display device 180 to display a display component indicating the inspection target area.
  • the inspection target storage unit 160 stores the three-dimensional coordinates of multiple inspection targets estimated by the inspection target position estimation unit 150 . Depending on the embodiment, additional information associated with the inspection object may be stored.
  • the inspection target storage unit 160 is a storage device that holds programs or various data (it may be located outside the inspection support device and connected via a network or the like to achieve the same purpose).
  • display control section 170 Based on the current camera image obtained from camera 110, the current position/orientation obtained from terminal position/orientation estimation section 120, and the three-dimensional coordinates of the inspection target held by inspection target storage section 160, display control section 170 , displays the AR on the screen. Since the position and orientation of the mobile terminal and the coordinates of the inspection object are in the same coordinate system of the three-dimensional space, it is sufficient to draw three-dimensional graphics representing the inspection object on the camera image. For drawing, for example, a three-dimensional graphics drawing API (Application Programming Interface) such as OpenGL (registered trademark) can be used.
  • a three-dimensional graphics drawing API Application Programming Interface
  • OpenGL registered trademark
  • FIG. 2 is a diagram showing an example of the mobile terminal 10 having the inspection support device 100 according to Embodiment 1 and an inspection object.
  • the equipment shown in FIG. 2 has a pipe 810, a tank 830, and valves 821 to 825 provided in the pipe 810 and subject to inspection. Further, near the valves 821-825, there are provided nameplates 821a-825a on which the names and identification numbers of inspection objects are written.
  • the equipment shown in FIG. 2 is an example of equipment to which the inspection support method according to Embodiment 1 can be applied, and the inspection support method can also be applied to other equipment.
  • the mobile terminal 10 detects the valve to be inspected, and displays the valve to be inspected on the display device 180 (for example, frame, highlighting, etc.).
  • FIG. 3 is a diagram showing a display example of the display device 180 of the mobile terminal 10 having the inspection support device 100 according to the first embodiment.
  • the user 800 simply points the camera 110 of the mobile terminal 10 toward the facility, and the mobile terminal 10 detects the valves 821 to 825 to be inspected, and displays the valves 821 to 825 to be inspected on the display device 180 (for example, frames of inspection target areas 821b to 825b, etc.).
  • FIG. 4 is a diagram showing an example of the hardware configuration of the inspection support device 100 and the mobile terminal 10 according to Embodiment 1.
  • the mobile terminal 10 is, for example, a smart phone, a tablet computer, or a personal computer (PC).
  • the mobile terminal 10 has an inspection support device 100 , a storage section 103 , a communication section 104 , a camera 110 , an inertial sensor 210 , a display device 180 and an input section 190 .
  • the inspection support device 100 has a processor 101 and a memory 102 .
  • Inertial sensors 210 include, for example, one or more of acceleration sensors, angular velocity sensors, and geomagnetic sensors.
  • the mobile terminal 10 is the inspection support system.
  • the inspection support device 100 has a processor 101 such as a CPU (Central Processing Unit) and a memory 102 that is a volatile storage device.
  • the memory 102 is, for example, a volatile semiconductor memory such as a RAM (Random Access Memory).
  • the inspection support device 100 may have a nonvolatile storage device such as a hard disk drive (HDD) or solid state drive (SSD).
  • HDD hard disk drive
  • SSD solid state drive
  • the processing circuitry may be dedicated hardware or processor 101 executing a program stored in memory 102 .
  • the processor 101 may be any of a processing device, an arithmetic device, a microprocessor, a microcomputer, and a DSP (Digital Signal Processor).
  • the processing circuit may be, for example, a single circuit, a composite circuit, a programmed processor, a parallel programmed processor, an ASIC (Application Specific Integrated Circuit), an FPGA (Field-Programmable Gate Array) ), or a combination of any of these.
  • ASIC Application Specific Integrated Circuit
  • FPGA Field-Programmable Gate Array
  • the inspection support program is realized by software, firmware, or a combination of software and firmware.
  • Software and firmware are written as programs and stored in memory 102 .
  • the processor 101 reads out and executes the inspection support program stored in the memory 102, thereby implementing the functions of the units shown in FIG.
  • inspection support device 100 may be partially realized by dedicated hardware and partially realized by software or firmware.
  • processing circuitry may implement each of the functions described above in hardware, software, firmware, or any combination thereof.
  • FIG. 5 is a diagram showing another example of the hardware configuration of the inspection support device 100a and the mobile terminal 10a according to the first embodiment.
  • the inspection support device 100a is a device capable of communicating with the mobile terminal 10a.
  • the mobile terminal 10a is, for example, a smart phone, a tablet computer, or a PC.
  • Portable terminal 10 a has processor 105 , memory 106 , storage unit 107 , communication unit 108 , camera 110 , inertial sensor 210 , display device 180 , and input unit 190 .
  • the inspection support device 100 a has a processor 101 , a memory 102 , a storage section 103 and a communication section 104 .
  • the inspection support device 100a has a processor 101 such as a CPU and a memory 102.
  • Memory 102 is, for example, a volatile semiconductor memory such as RAM.
  • the inspection support device 100a may have a non-volatile storage device such as an HDD or SSD.
  • the combination of the mobile terminal 10a and the inspection support device 100a is the inspection support system.
  • FIG. 6 is a flowchart showing the terminal position/orientation estimation processing S11 in the terminal position/orientation estimation thread by the inspection support apparatus 100 according to the first embodiment.
  • the terminal position/orientation estimation unit 120 acquires the current camera image from the camera 110 and the current sensor information from the sensor, estimates the current position/orientation of the mobile terminal, and provides the inspection target position estimation unit 150 and the display control unit 170 Notify the position and orientation.
  • FIG. 7 is a flowchart showing an inspection target acquisition thread by the inspection support device 100 according to the first embodiment.
  • the inspection object detection unit 130 acquires the current camera image from the camera 110 and detects zero or more inspection object areas from the camera image (step S21).
  • the inspection target tracking unit 140 tracks the inspection target area between the camera images in time series (step S22).
  • the inspection target position estimation unit 150 estimates the position (ie, three-dimensional coordinates) of the inspection target region based on the current position and orientation of the mobile terminal 10 and the tracked inspection target region (step S24).
  • the coordinates estimated above are compared with the coordinates of each inspection object stored in the inspection object storage unit 160, and if the coordinates are the same or the distance between the coordinates is within a certain value ( Steps S25, S26), it is determined that the inspection object has already been added to the inspection object storage unit 160, the loop processing returns to the beginning, and in other cases, the inspection object is added to the inspection object storage unit 160. (Step S27).
  • the display control unit 170 acquires the current camera image from the camera 110, the current position and orientation of the mobile terminal from the terminal position/orientation estimation unit 120, and a plurality of inspection targets from the inspection target storage unit 160, Display the AR on the screen.
  • the inspection support device 100 and the mobile terminal 10 according to Embodiment 1 there is no need to perform operations such as attaching a marker to an inspection target and creating a three-dimensional model in advance.
  • the object to be inspected can be displayed by AR without creating location information of the object to be inspected in advance. Since the AR is displayed by estimating the three-dimensional coordinates of the inspection target, even if the inspection target is no longer detected or is hidden behind something, the AR indicating the position of the inspection target can be continuously displayed. can.
  • FIG. 8 is a functional block diagram schematically showing configurations of inspection support device 200 and mobile terminal 20 according to Embodiment 2. As shown in FIG. 8, the same or corresponding components as those shown in FIG. 1 are given the same reference numerals as those shown in FIG.
  • Inspection support apparatus 200 is an inspection support apparatus in that it includes character recognition unit 220 and recognition target acquisition unit 230, and terminal position/orientation estimation unit 120 estimates the position and orientation using the output of inertial sensor 210. Different from 100.
  • FIG. 9 is a diagram showing a display example of a camera image of the mobile terminal 20 according to Embodiment 2 (at the time of reading a character string).
  • the character recognition unit 220 recognizes, from the camera image acquired by the camera 110, the identification number, which is the character string of the nameplate 821a that identifies the valve 821 to be inspected. Recognition processing is executed when instructed by the user or is always executed.
  • the recognition target acquisition unit 230 stores an inspection target corresponding to the inspection target with the identification number recognized by the character recognition unit 220.
  • the inspection target stored by the unit 160 is acquired. That is, among the three-dimensional coordinates of a plurality of inspection objects stored in the inspection object storage unit 160, an inspection object having three-dimensional coordinates corresponding to the inspection object recognized by the character recognition unit 220 is searched. Specifically, for example, the recognition target acquisition unit 230 searches the inspection target storage unit 160 for an inspection target having three-dimensional coordinates closest to the current position of the mobile terminal 20, and sets it as the current recognition target.
  • the recognition target acquiring unit 230 searches for the inspection target having the closest three-dimensional coordinates from a position that is an arbitrary distance (for example, 30 cm) in the viewing direction of the camera 110 from the current position of the mobile terminal 20, and performs the current recognition. set to target.
  • the recognition target acquisition unit 230 may add the information “checked” to the inspection target stored in the inspection target storage unit 160 to display the inspection target differently from the uninspected inspection target. In this manner, the inspection target storage unit 160 holds information indicating whether or not each inspection target has been inspected in addition to the information described in the first embodiment.
  • FIG. 10 is a diagram showing a display example of the display device 180 of the mobile terminal 20 according to the second embodiment.
  • the display control unit 170 changes the display method of the inspection object according to whether the inspection object recognized by the recognition object acquisition unit 230 has been inspected.
  • the display is performed using AR, for example.
  • the color of the frame of the AR image that surrounds the inspection target that has already been inspected and the color of the frame of the AR image that surrounds the inspection target that has not yet been inspected are set to different colors.
  • FIG. 10 shows an example in which an AR image is displayed superimposed on a camera image.
  • an AR image can be displayed superimposed on a real landscape.
  • AR images may be projected onto real installations (eg, pipes, valves).
  • the mobile terminal 20 has an application for forming an illustration figure simulating a camera image
  • the AR image may be superimposed on the illustration simulating the camera image.
  • the identification number of the inspection object is recognized by character recognition, and the inspection object storage unit 160 determines that the inspection object has been inspected.
  • the inspection object storage unit 160 determines that the inspection object has been inspected.
  • the second embodiment is the same as the first embodiment.
  • FIG. 11 is a functional block diagram schematically showing configurations of inspection support device 300 and mobile terminal 30 according to the third embodiment. 11, the same or corresponding components as those shown in FIG. 8 are given the same reference numerals as those shown in FIG. Inspection support apparatus 300 according to Embodiment 3 differs from inspection support apparatus 200 according to Embodiment 2 in terms of the function of recognition target acquisition section 230a.
  • the inspection target position estimating unit 150 estimates the coordinates of the inspection target region in the three-dimensional coordinate system based on the position and orientation of the mobile terminal 30 and the inspection target regions in at least two of the plurality of camera images. and stored in the inspection target storage unit 160 .
  • the display control unit 170 causes the display device 180 to display display components (for example, AR display) and recognized characters.
  • the recognition target acquisition unit 230a also stores the identification number recognized by the character recognition unit 220 when rewriting the inspection target stored in the inspection target storage unit 160 to "inspected".
  • the inspection target storage unit 160 also stores the identification number of each inspection target. If it has not been inspected, the identification number shall be unknown.
  • the display control unit 170 also displays the identification number in AR if the identification number of the inspection target is known.
  • FIG. 12 is a diagram showing a display example of the display device 180 of the mobile terminal 30 according to the third embodiment.
  • FIG. 12 displays an identification number 821c around (that is, in the vicinity of) the valve 821 to be inspected.
  • the inspection support device 300 and the mobile terminal 30 according to the third embodiment by displaying the identification number in the vicinity of the recognized inspection object by the AR image, easier to distinguish.
  • Embodiment 3 is the same as Embodiment 1 or 2.
  • FIG. 13 is a functional block diagram schematically showing configurations of inspection support device 400 and portable terminal 40 according to the fourth embodiment. 13, the same or corresponding components as those shown in FIG. 11 are given the same reference numerals as those shown in FIG. Inspection support apparatus 400 according to Embodiment 4 differs from inspection support apparatus 300 according to Embodiment 3 in that it includes terminal position/posture storage section 410 and terminal position/posture recovery section 420 .
  • the terminal position/orientation estimation unit 120 associates the camera image with the position/orientation of the mobile terminal 40 and stores them in the terminal position/orientation storage unit 410 .
  • the terminal position/orientation restoring unit 420 estimates the terminal position/orientation using the position/orientation associated with the camera image closest to the current camera image among the camera images stored in the terminal position/orientation estimation unit 120 as the current position/orientation. 120.
  • the terminal position/orientation storage unit 410 periodically (for example, every 10 seconds) stores the relationship between the current camera image and the current position/orientation of the mobile terminal estimated by the terminal position/orientation estimation unit 120 . A plurality of such relationships are held.
  • the camera image may not be raw data of the image, but may be processed/processed data as long as it is in a format that enables matching between images, which will be described later.
  • the data may be in formats such as SIFT (Scale-invariant feature transform), SURF (Speed-Upped Robust Feature), ORB (Oriented FAST and Rotated BRIEF), and KAZE, which are local features.
  • the terminal position/orientation recovery unit 420 matches the current camera image with each camera image stored in the terminal position/orientation storage unit 410 when the inspection support device is activated, and returns the terminal position/orientation closest to the current camera image.
  • a camera image stored in the storage unit 410 is specified.
  • the terminal position/orientation estimating unit 120 is initialized with the position/orientation of the mobile terminal linked to the camera image as the current position/orientation of the mobile terminal.
  • the camera image and the position and orientation of the mobile terminal 40 are periodically stored, and when the mobile terminal 40 is activated, the current
  • the previous camera image with the camera image stored in the past the previous three-dimensional coordinate system can be inherited and the inspection work can be restarted. Therefore, the AR display can be resumed immediately without the need to detect and estimate the position of the object to be inspected again.
  • Embodiment 4 is the same as any of Embodiments 1 to 3.
  • FIG. 14 is a functional block diagram schematically showing configurations of inspection support device 500 and mobile terminal 50 according to Embodiment 5. As shown in FIG. 14, the same or corresponding components as those shown in FIG. 13 are given the same reference numerals as those shown in FIG. Inspection support device 500 according to Embodiment 5 differs from inspection support device 400 according to Embodiment 4 in that inspection state initialization section 510 is provided.
  • the inspection state initialization unit 510 initializes some or all of the inspections held by the inspection target storage unit 160 when the inspection support device 500 is activated or when specified by the user via the input unit (190 in FIG. 4). Information on the inspection status of the target is set to uninspected.
  • a new inspection can be started using the inspection support device 500 and the mobile terminal 50 according to the fifth embodiment.
  • the three-dimensional coordinates used to represent the coordinates of the inspection target have already been obtained once, it may not be necessary to detect the inspection target and estimate the coordinates again. Inspection can be started in the displayed state.
  • the identification number is also known, it is possible to search for an uninspected object to be inspected based on the identification number.
  • Embodiment 5 is the same as any of Embodiments 1 to 4.
  • FIG. 15 is a functional block diagram schematically showing configurations of inspection support device 600 and mobile terminal 60 according to Embodiment 6. As shown in FIG. 15, the same or corresponding components as those shown in FIG. 14 are given the same reference numerals as those shown in FIG.
  • the inspection support device 600 according to the sixth embodiment differs from the inspection support device 500 according to the fifth embodiment in that it has an inspection procedure storage unit 610 and an inspection procedure management unit 620 .
  • the inspection procedure storage unit 610 stores information indicating an inspection procedure created in advance.
  • the information indicating the inspection procedure is information indicating the order of inspection of a plurality of inspection targets, and is specifically a list of arranged identification numbers.
  • the inspection procedure management unit 620 manages the current inspection procedure status. If the identification number is read by the character recognition unit 220 and the identification number is the identification number of the current inspection procedure, the inspection procedure is advanced by one. Information on the current inspection procedure is provided to the display control unit 670 . That is, the inspection procedure management unit 620 provides information to the display control unit 670 based on the order of inspection of the inspection objects stored in the inspection procedure storage unit 610 and the recognized characters, and displays the next inspection object. is displayed on the display device 180 .
  • the display control unit 670 displays the inspection target. Display the target AR so that it stands out compared to other inspection targets. For example, when it is the identification number of an inspection object currently being inspected managed by the inspection procedure management unit 620, the display control unit 670 makes the AR display for the inspection object stand out compared to other inspection objects. , change the color of the AR display, or blink the display part indicating the inspection target.
  • FIG. 16 is a diagram showing a display example of the display device 180 of the mobile terminal 60 according to the sixth embodiment.
  • FIG. 16 shows that the valve 823 is to be inspected next.
  • the display method of the next inspection target is not limited to that of FIG.
  • the inspection support device 600 and the mobile terminal 60 according to Embodiment 6 it is possible to display the position of the inspection target in AR, so that the inspection can be performed intuitively according to the procedure. .
  • Embodiment 6 is the same as any of Embodiments 1 to 5.
  • FIG. 17 is a functional block diagram schematically showing configurations of inspection support device 700 and mobile terminal 70 according to Embodiment 7. As shown in FIG. 17, the same or corresponding components as those shown in FIG. 14 are given the same reference numerals as those shown in FIG. Inspection support device 700 according to Embodiment 7 differs from inspection support device 500 according to Embodiment 5 in that it includes an identification number confirmation unit 710 .
  • Identification number confirmation unit 710 confirms that the identification number recognized this time by character recognition unit 220 is the identification number for the same inspection object already stored in inspection object storage unit 160 (that is, the identification number recognized in the past). It further has an identification number confirming unit for displaying a warning on the screen of the display device 180 when the numbers do not match. For example, when the character recognition unit 220 and the recognition target acquisition unit 230 change the inspection status information of the inspection target stored in the inspection target storage unit 160 from "unchecked" to "checked", the inspection target storage unit 160 has already stored an identification number, and the identification number is different from the identification number recognized this time, the identification number confirmation unit 710 displays a warning on the display device. 180 screen.
  • FIG. 18 is a diagram showing a display example of the display device 180 of the mobile terminal 70 according to the seventh embodiment.
  • identification number confirmation unit 710 issues a warning indicating that the identification number recognized this time does not match the identification number for the same inspection object already stored in inspection object storage unit 160 . It is displayed on the screen of the display device 180 .
  • the warning is not limited to that shown in FIG. 18, and may be other text, video, audio, and the like.
  • the identification number recognized in the past and the identification number recognized this time are different for the same inspection object. It is possible to inform the user of an abnormal situation that there is
  • the abnormal situation is, for example, a change in the arrangement of inspection targets in the field equipment, a change in the inspection target nameplate, or the like.
  • Embodiment 7 is the same as any of Embodiments 1 to 6.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Telephone Function (AREA)

Abstract

This inspection assistance device (100) has: a terminal position/orientation estimation unit (120) that estimates the position and orientation of a portable terminal (10) in a three-dimensional coordinate system; an inspection object detection unit (130) that detects an inspection object from a camera image photographed by a camera (110); an inspection object tracking unit (140) that tracks inspection object regions (821-825), which indicate the inspection object detected in the camera image, in time series in the order of time of photographing, and that identifies the same inspection object regions (821b-825b) in a plurality of camera images arranged in time series; an inspection object position estimation unit (150) that estimates the coordinates of the inspection object regions in the three-dimensional coordinate system on the basis of the position and orientation of the portable terminal (10) and inspection object regions in at least two camera images among the plurality of camera images; and a display control unit (170) that causes a display device (180) to display a display component for making the inspection object regions visible.

Description

点検支援装置、点検支援システム、点検支援方法、及び点検支援プログラムInspection support device, inspection support system, inspection support method, and inspection support program
 本開示は、点検支援装置、点検支援システム、点検支援方法、及び点検支援プログラムに関する。 The present disclosure relates to an inspection support device, an inspection support system, an inspection support method, and an inspection support program.
 スマートフォン及びタブレット型コンピュータなどの携帯端末により、拡張現実(AR:Augmented Reality)技術を用いて、インフラ施設などにおける設備の点検を支援する技術の提案がある。例えば、特許文献1を参照。この技術が適用された携帯端末は、基準マーカが備えられた点検対象を検出し、検出された点検対象の登録情報を管理データベースから取得し、取得した情報をカメラ映像の上に重畳表示することができる。 There is a proposal for a technology that supports the inspection of infrastructure facilities using mobile terminals such as smartphones and tablet computers, using augmented reality (AR) technology. See, for example, US Pat. A mobile terminal to which this technology is applied detects an inspection target equipped with a reference marker, acquires registration information of the detected inspection target from the management database, and displays the acquired information superimposed on the camera image. can be done.
特開2020-80147号公報(例えば、段落0012~0037)JP 2020-80147 (for example, paragraphs 0012 to 0037)
 しかしながら、上記従来の技術では、ユーザは、事前に点検対象に基準マーカを取り付ける、及び事前に点検対象の位置情報をデータベースに入力する、などの面倒な準備作業を行う必要があった。 However, with the above conventional technology, the user had to perform troublesome preparatory work such as attaching reference markers to the inspection target in advance and inputting the position information of the inspection target into the database in advance.
 本開示は、準備作業を必要とせずに点検作業を支援することを可能にする点検支援装置、点検支援システム、点検支援方法、及び点検支援プログラムを提供することを目的とする。 An object of the present disclosure is to provide an inspection support device, an inspection support system, an inspection support method, and an inspection support program that can support inspection work without requiring preparatory work.
 本開示の点検支援装置は、携帯端末に備えられたカメラにより撮影されたカメラ画像と前記携帯端末に備えられた慣性センサにより得られた検出信号とのうちの少なくとも一方に基づいて、3次元座標系における前記携帯端末の位置姿勢を推定する端末位置姿勢推定部と、前記カメラにより撮影されたカメラ画像から、点検対象を検出する点検対象検出部と、前記カメラ画像において検出された前記点検対象を示す点検対象領域を撮影時刻順に時系列で追跡して、時系列に並ぶ複数枚のカメラ画像において同じ点検対象領域を特定する点検対象追跡部と、前記携帯端末の前記位置姿勢と前記複数枚のカメラ画像のうちの少なくとも2枚のカメラ画像における前記点検対象領域とに基づいて、前記点検対象領域の前記3次元座標系における座標を推定する点検対象位置推定部と、前記点検対象領域を視認可能にする表示部品を表示装置に表示させる表示制御部と、を有することを特徴とする。 An inspection support device according to the present disclosure, based on at least one of a camera image captured by a camera provided in a mobile terminal and a detection signal obtained by an inertial sensor provided in the mobile terminal, determines three-dimensional coordinates a terminal position/orientation estimation unit that estimates the position and orientation of the mobile terminal in the system; an inspection object detection unit that detects an inspection object from a camera image captured by the camera; an inspection target tracking unit that tracks the indicated inspection target area in chronological order of photographing time and identifies the same inspection target area in a plurality of camera images arranged in time series; an inspection target position estimating unit for estimating the coordinates of the inspection target region in the three-dimensional coordinate system based on the inspection target region in at least two camera images out of the camera images; and a display control unit that causes a display device to display a display component to be used.
 本開示の点検支援方法は、点検支援装置によって実行される方法であって、携帯端末に備えられたカメラにより撮影されたカメラ画像と前記携帯端末に備えられた慣性センサにより得られた検出信号とのうちの少なくとも一方に基づいて、3次元座標系における前記携帯端末の位置姿勢を推定するステップと、前記カメラにより撮影されたカメラ画像から、点検対象を検出するステップと、前記カメラ画像において検出された前記点検対象を示す点検対象領域を撮影時刻順に時系列で追跡して、時系列に並ぶ複数枚のカメラ画像において同じ点検対象領域を特定するステップと、前記携帯端末の前記位置姿勢と前記複数枚のカメラ画像のうちの少なくとも2枚のカメラ画像における前記点検対象領域とに基づいて、前記点検対象領域の前記3次元座標系における座標を推定するステップと、前記点検対象領域を視認可能にする表示部品を表示装置に表示させるステップと、を有することを特徴とする。 An inspection support method of the present disclosure is a method executed by an inspection support device, and includes a camera image captured by a camera provided in a mobile terminal and a detection signal obtained by an inertial sensor provided in the mobile terminal. estimating the position and orientation of the mobile terminal in a three-dimensional coordinate system based on at least one of; detecting an inspection target from a camera image captured by the camera; a step of tracking the inspection target area indicating the inspection target in chronological order of photographing time, and identifying the same inspection target area in a plurality of camera images arranged in time series; estimating the coordinates of the inspection target area in the three-dimensional coordinate system based on the inspection target area in at least two camera images out of the camera images; and making the inspection target area visible. and a step of displaying the display component on a display device.
 本開示によれば、準備作業を必要とせずに点検作業を支援することができる。 According to the present disclosure, inspection work can be supported without the need for preparatory work.
実施の形態1に係る点検支援装置及び携帯端末の構成を概略的に示す機能ブロック図である。1 is a functional block diagram schematically showing configurations of an inspection support device and a portable terminal according to Embodiment 1; FIG. 実施の形態1に係る端末装置と点検対象の例を示す図である。FIG. 2 is a diagram showing an example of a terminal device and an inspection target according to Embodiment 1; FIG. 実施の形態1に係る端末装置の表示装置の表示例を示す図である。4 is a diagram showing a display example of the display device of the terminal device according to Embodiment 1; FIG. 実施の形態1に係る点検支援装置及び携帯端末のハードウェア構成の例を示す図である。2 is a diagram illustrating an example of hardware configurations of an inspection support device and a mobile terminal according to Embodiment 1; FIG. 実施の形態1に係る点検支援装置及び携帯端末のハードウェア構成の他の例を示す図である。3 is a diagram showing another example of the hardware configuration of the inspection support device and the mobile terminal according to Embodiment 1; FIG. 実施の形態1に係る点検支援装置による端末位置姿勢推定スレッドを示すフローチャートである。7 is a flowchart showing a terminal position/orientation estimation thread by the inspection support device according to the first embodiment; 実施の形態1に係る点検支援装置による点検対象取得スレッドを示すフローチャートである。7 is a flow chart showing an inspection target acquisition thread by the inspection support device according to the first embodiment; 実施の形態2に係る点検支援装置及び携帯端末の構成を概略的に示す機能ブロック図である。FIG. 10 is a functional block diagram schematically showing configurations of an inspection support device and a portable terminal according to Embodiment 2; 実施の形態2に係る端末装置のカメラ画像の表示例(文字列読取時)を示す図である。FIG. 10 is a diagram showing a display example of a camera image of the terminal device according to Embodiment 2 (at the time of reading a character string); 実施の形態2に係る端末装置の表示装置の表示例を示す図である。FIG. 11 is a diagram showing a display example of a display device of a terminal device according to Embodiment 2; 実施の形態3に係る点検支援装置及び携帯端末の構成を概略的に示す機能ブロック図である。FIG. 11 is a functional block diagram schematically showing the configuration of an inspection support device and a portable terminal according to Embodiment 3; 実施の形態3に係る端末装置の表示装置の表示例を示す図である。FIG. 11 is a diagram showing a display example of a display device of a terminal device according to Embodiment 3; 実施の形態4に係る点検支援装置及び携帯端末の構成を概略的に示す機能ブロック図である。FIG. 11 is a functional block diagram schematically showing the configuration of an inspection support device and a portable terminal according to Embodiment 4; 実施の形態5に係る点検支援装置及び携帯端末の構成を概略的に示す機能ブロック図である。FIG. 11 is a functional block diagram schematically showing the configuration of an inspection support device and a portable terminal according to Embodiment 5; 実施の形態6に係る点検支援装置及び携帯端末の構成を概略的に示す機能ブロック図である。FIG. 12 is a functional block diagram schematically showing the configuration of an inspection support device and a portable terminal according to Embodiment 6; 実施の形態6に係る端末装置の表示装置の表示例を示す図である。FIG. 13 is a diagram showing a display example of a display device of a terminal device according to Embodiment 6; 実施の形態7に係る点検支援装置及び携帯端末の構成を概略的に示す機能ブロック図である。FIG. 14 is a functional block diagram schematically showing the configuration of an inspection support device and a portable terminal according to Embodiment 7; 実施の形態7に係る端末装置の表示装置の表示例を示す図である。FIG. 12 is a diagram showing a display example of a display device of a terminal device according to Embodiment 7;
 以下に、実施の形態に係る点検支援装置、点検支援システム、点検支援方法、及び点検支援プログラムを、図面を参照しながら説明する。以下の実施の形態は、例にすぎず、実施の形態を適宜組み合わせること及び各実施の形態を適宜変更することが可能である。 The inspection support device, inspection support system, inspection support method, and inspection support program according to the embodiment will be described below with reference to the drawings. The following embodiments are merely examples, and the embodiments can be combined as appropriate and each embodiment can be modified as appropriate.
実施の形態1.
 図1は、実施の形態1に係る点検支援装置100及び携帯端末10の構成を概略的に示す機能ブロック図である。点検支援装置100は、実施の形態1に係る点検支援方法を実行することができる装置である。点検支援装置100は、点検支援プログラムを実行することにより点検支援方法を実行することができる。
Embodiment 1.
FIG. 1 is a functional block diagram schematically showing configurations of an inspection support device 100 and a mobile terminal 10 according to Embodiment 1. As shown in FIG. The inspection support device 100 is a device capable of executing the inspection support method according to the first embodiment. The inspection support device 100 can execute an inspection support method by executing an inspection support program.
 図1に示されるように、携帯端末10は、点検支援装置100と、カメラ110と、表示装置180とを有している。カメラ110は、例えば、RGBカメラである。カメラ110は、ステレオカメラであってもよい。表示装置180は、例えば、液晶表示装置などの、表示装置である。表示装置180は、表示装置としての機能とユーザ操作部である入力部としての機能とを備えたタッチパネルであってもよい。また、携帯端末10は、慣性センサ(後述の図4で説明される)を備えてもよい。 As shown in FIG. 1, the mobile terminal 10 has an inspection support device 100, a camera 110, and a display device 180. Camera 110 is, for example, an RGB camera. Camera 110 may be a stereo camera. The display device 180 is, for example, a display device such as a liquid crystal display device. The display device 180 may be a touch panel that has a function as a display device and a function as an input unit that is a user operation unit. The mobile terminal 10 may also include an inertial sensor (described later with reference to FIG. 4).
 図1に示されるように、点検支援装置100は、端末位置姿勢推定部120と、点検対象検出部130と、点検対象追跡部140と、点検対象位置推定部150と、表示装置180に画像を表示させる表示制御部170とを有している。点検支援装置100は、情報を記憶する記憶装置である点検対象記憶部160を有してもよい。 As shown in FIG. 1 , inspection support device 100 includes terminal position/orientation estimation unit 120 , inspection object detection unit 130 , inspection object tracking unit 140 , inspection object position estimation unit 150 , and display device 180 . and a display control unit 170 for displaying. The inspection support device 100 may have an inspection target storage unit 160 that is a storage device that stores information.
 端末位置姿勢推定部120は、カメラ110で撮影したカメラ画像から、世界空間を表す3次元座標系における携帯端末の位置姿勢(すなわち、位置及び姿勢)を推定する。この3次元座標系は、予め定められた位置(例えば、地球上における予め決められた経緯度又は予め決められた建築物の点)を原点とする座標系である必要はない。3次元座標系は、予め定められた座標系である必要はなく、例えば、点検支援装置100の動作開始時(例えば、携帯端末10の電源投入時又は点検支援プログラムの実行アプリの起動時)における携帯端末10の位置及び姿勢に基づいて決定すればよい。つまり、端末位置姿勢推定部120は、携帯端末の起動時(すなわち、装置起動時)又は点検支援プログラムの起動時(すなわち、アプリ起動時)における携帯端末の位置及び姿勢に基づいて、原点及び3次元座標系を設定してもよい。したがって、起動時(すなわち、装置起動時又はアプリ起動時)における携帯端末の位置姿勢から、携帯端末の位置姿勢の変化を積分していくことで、起動時の位置を原点とした3次元座標系における、各時刻における携帯端末の位置姿勢を推定することができる。推定方法としては、カメラ画像中の特徴量の情報に基づいて各時刻における位置姿勢を計算する方法、加速度センサによる速度及び角速度センサによる回転などの情報に基づいて各時刻における位置姿勢を計算する方法がある。位置姿勢を推定する方法としては、既存の方法を使用することができる。 The terminal position and orientation estimation unit 120 estimates the position and orientation (that is, the position and orientation) of the mobile terminal in a three-dimensional coordinate system representing the world space from the camera image captured by the camera 110 . This three-dimensional coordinate system does not need to be a coordinate system whose origin is a predetermined position (for example, predetermined longitude and latitude on the earth or a predetermined building point). The three-dimensional coordinate system does not have to be a predetermined coordinate system. It may be determined based on the position and orientation of the mobile terminal 10 . That is, the terminal position/orientation estimating unit 120 calculates the position and orientation of the mobile terminal based on the position and orientation of the mobile terminal when the mobile terminal is activated (that is, when the device is activated) or when the inspection support program is activated (that is, when the application is activated). A dimensional coordinate system may be set. Therefore, by integrating the change in the position and orientation of the mobile terminal from the position and orientation of the mobile terminal at the time of startup (that is, at the time of device startup or application startup), a three-dimensional coordinate system with the position at the time of startup as the origin , the position and orientation of the mobile terminal at each time can be estimated. As an estimation method, there is a method of calculating the position and orientation at each time based on the information of the feature amount in the camera image, and a method of calculating the position and orientation at each time based on information such as the velocity from the acceleration sensor and the rotation from the angular velocity sensor. There is An existing method can be used as a method for estimating the position and orientation.
 携帯端末10が慣性センサ(図3に示される)を有する場合には、端末位置姿勢推定部120は、カメラ110により撮影されたカメラ画像と慣性センサにより得られた検出信号とのうちの少なくとも一方に基づいて、3次元座標系における携帯端末10の位置及び姿勢を推定してもよい。また、端末位置姿勢推定部120は、カメラ110とは異なるカメラで撮影されたカメラ画像と携帯端末10に備えられた慣性センサ210により得られた検出信号とのうちの少なくとも一方に基づいて、3次元座標系における前記携帯端末(10)の位置姿勢を推定してもよい。 If the mobile terminal 10 has an inertial sensor (shown in FIG. 3), the terminal position/orientation estimation unit 120 detects at least one of a camera image captured by the camera 110 and a detection signal obtained by the inertial sensor. , the position and orientation of the mobile terminal 10 in the three-dimensional coordinate system may be estimated. In addition, terminal position/orientation estimation section 120 performs 3 A position and orientation of the mobile terminal (10) in a dimensional coordinate system may be estimated.
 点検対象検出部130は、カメラ110により撮影されたカメラ画像から、点検対象(例えば、バルブ、計器、など)を検出する。 The inspection target detection unit 130 detects inspection targets (eg, valves, gauges, etc.) from the camera image captured by the camera 110 .
 点検対象追跡部140は、時系列に並ぶ複数枚のカメラ画像において検出された点検対象を示す点検対象領域を追跡し、時系列に並ぶ複数枚のカメラ画像において同じ点検対象領域を特定する。点検対象追跡部140は、点検対象検出部130によって検出された現在のカメラ画像における点検対象の領域と、前回の検出における点検対象の領域とを比較することで、カメラ画像上の点検対象領域の位置の変化を時系列で追跡する。点検対象検出部130は、カメラ画像単位(すなわち、カメラ画像のフレーム毎)で検出処理を行うため、時間の前後のカメラ画像で検出された点検対象領域との関係性は不明である。点検対象追跡部140の処理により、時間の前後のカメラ画像で検出された点検対象領域の関係性を把握する。点検対象追跡部140は、例えば、検出された点検対象領域の矩形の一部が、時間の前後のカメラ画像間で重なっている場合、同一の点検対象を示していると判断する。あるいは、点検対象追跡部140は、オプティカルフローと呼ばれる手法を用いて、カメラ画像間の画素の移動方向と移動量を算出し、前後の点検対象領域と画素の移動とを比較することで、点検対象領域の関係性を特定することができる。 The inspection target tracking unit 140 tracks the inspection target area indicating the inspection target detected in the multiple camera images arranged in time series, and identifies the same inspection target area in the multiple camera images arranged in time series. The inspection target tracking unit 140 compares the inspection target region in the current camera image detected by the inspection target detection unit 130 with the inspection target region in the previous detection, thereby determining the inspection target region on the camera image. Track position changes over time. Since the inspection target detection unit 130 performs detection processing for each camera image (that is, each frame of the camera image), the relationship between the inspection target areas detected in the camera images before and after the time is unknown. Through the processing of the inspection object tracking unit 140, the relationship between the inspection object areas detected in the camera images before and after the time is grasped. For example, the inspection target tracking unit 140 determines that the same inspection target is indicated when a part of the detected inspection target region rectangle overlaps between the camera images before and after the time. Alternatively, the inspection target tracking unit 140 uses a technique called optical flow to calculate the movement direction and amount of movement of pixels between camera images, and compares the previous and subsequent inspection target areas with the movement of pixels, thereby performing an inspection. Able to identify relationships between regions of interest.
 点検対象位置推定部150は、携帯端末10の位置及び姿勢と複数枚のカメラ画像のうちの少なくとも2枚のカメラ画像における点検対象領域とに基づいて、点検対象領域の3次元座標系における座標を推定する。 The inspection target position estimation unit 150 calculates the coordinates of the inspection target region in the three-dimensional coordinate system based on the position and orientation of the mobile terminal 10 and the inspection target regions in at least two camera images out of the plurality of camera images. presume.
 点検対象記憶部160は、点検対象を示す点検対象領域の座標を記憶する。表示制御部170は、点検対象領域を示す表示部品を表示装置180に表示させる。 The inspection target storage unit 160 stores the coordinates of the inspection target area indicating the inspection target. The display control unit 170 causes the display device 180 to display a display component indicating the inspection target area.
 点検対象記憶部160は、点検対象位置推定部150によって推定された複数の点検対象の3次元座標を記憶する。実施の形態に応じて、点検対象に紐付けた追加の情報を記憶してもよい。点検対象記憶部160は、プログラム又は各種データを保持する記憶装置である(点検支援装置の外部にあって、ネットワーク等で接続することによって、同等の目的を実現してもよい)。 The inspection target storage unit 160 stores the three-dimensional coordinates of multiple inspection targets estimated by the inspection target position estimation unit 150 . Depending on the embodiment, additional information associated with the inspection object may be stored. The inspection target storage unit 160 is a storage device that holds programs or various data (it may be located outside the inspection support device and connected via a network or the like to achieve the same purpose).
 表示制御部170は、カメラ110から得た現在のカメラ画像と、端末位置姿勢推定部120から得た現在の位置姿勢と、点検対象記憶部160が保持する点検対象の3次元座標とに基づいて、画面にARを表示する。携帯端末の位置姿勢と、点検対象の座標は、同じ3次元空間の座標系であるから、カメラ画像上に点検対象を表す3次元グラフィックスを描画すればよい。描画には、例えば、OpenGL(登録商標)のような3次元グラフィックス描画API(Application Programming Interface)を利用することができる。 Based on the current camera image obtained from camera 110, the current position/orientation obtained from terminal position/orientation estimation section 120, and the three-dimensional coordinates of the inspection target held by inspection target storage section 160, display control section 170 , displays the AR on the screen. Since the position and orientation of the mobile terminal and the coordinates of the inspection object are in the same coordinate system of the three-dimensional space, it is sufficient to draw three-dimensional graphics representing the inspection object on the camera image. For drawing, for example, a three-dimensional graphics drawing API (Application Programming Interface) such as OpenGL (registered trademark) can be used.
 図2は、実施の形態1に係る点検支援装置100を有する携帯端末10と点検対象の例を示す図である。図2に示される設備は、配管810と、タンク830と、配管810に備えられた点検対象としてのバルブ821~825とを有している。また、バルブ821~825の近くには、点検対象の名称及び識別番号が記載された銘板821a~825aが備えられている。図2に示される設備は、実施の形態1に係る点検支援方法を適用可能な設備の一例であり、他の設備にも点検支援方法を適用可能である。ユーザ800が携帯端末10を持って移動すると、携帯端末10は、点検対象のバルブを検出し、表示装置180に点検対象のバルブを示す表示(例えば、枠、強調表示、など)を行う。 FIG. 2 is a diagram showing an example of the mobile terminal 10 having the inspection support device 100 according to Embodiment 1 and an inspection object. The equipment shown in FIG. 2 has a pipe 810, a tank 830, and valves 821 to 825 provided in the pipe 810 and subject to inspection. Further, near the valves 821-825, there are provided nameplates 821a-825a on which the names and identification numbers of inspection objects are written. The equipment shown in FIG. 2 is an example of equipment to which the inspection support method according to Embodiment 1 can be applied, and the inspection support method can also be applied to other equipment. When the user 800 moves with the mobile terminal 10, the mobile terminal 10 detects the valve to be inspected, and displays the valve to be inspected on the display device 180 (for example, frame, highlighting, etc.).
 図3は、実施の形態1に係る点検支援装置100を有する携帯端末10の表示装置180の表示例を示す図である。ユーザ800は、携帯端末10のカメラ110を設備に向けるだけで、携帯端末10は、点検対象のバルブ821~825を検出し、表示装置180に点検対象のバルブ821~825を示す表示(例えば、点検対象領域821b~825bの枠、など)を行う。 FIG. 3 is a diagram showing a display example of the display device 180 of the mobile terminal 10 having the inspection support device 100 according to the first embodiment. The user 800 simply points the camera 110 of the mobile terminal 10 toward the facility, and the mobile terminal 10 detects the valves 821 to 825 to be inspected, and displays the valves 821 to 825 to be inspected on the display device 180 (for example, frames of inspection target areas 821b to 825b, etc.).
 図4は、実施の形態1に係る点検支援装置100及び携帯端末10のハードウェア構成の例を示す図である。携帯端末10は、例えば、スマートフォン、タブレット型コンピュータ、又はパーソナルコンピュータ(PC)、などである。携帯端末10は、点検支援装置100と、記憶部103と、通信部104と、カメラ110と、慣性センサ210と、表示装置180と、入力部190とを有している。点検支援装置100は、プロセッサ101と、メモリ102とを有している。慣性センサ210は、例えば、加速度センサ、角速度センサ、及び地磁気センサのうちの1つ以上を含む。図4の例では、携帯端末10が点検支援システムである。 FIG. 4 is a diagram showing an example of the hardware configuration of the inspection support device 100 and the mobile terminal 10 according to Embodiment 1. As shown in FIG. The mobile terminal 10 is, for example, a smart phone, a tablet computer, or a personal computer (PC). The mobile terminal 10 has an inspection support device 100 , a storage section 103 , a communication section 104 , a camera 110 , an inertial sensor 210 , a display device 180 and an input section 190 . The inspection support device 100 has a processor 101 and a memory 102 . Inertial sensors 210 include, for example, one or more of acceleration sensors, angular velocity sensors, and geomagnetic sensors. In the example of FIG. 4, the mobile terminal 10 is the inspection support system.
 図4に示されるように、点検支援装置100は、CPU(Central Processing Unit)などのプロセッサ101と揮発性の記憶装置であるメモリ102とを有する。メモリ102は、例えば、RAM(Random Access Memory)などの、揮発性の半導体メモリである。図1に示されるように、点検支援装置100は、ハードディスクドライブ(HDD)又はソリッドステートドライブ(SSD)などの不揮発性の記憶装置を有してもよい。 As shown in FIG. 4, the inspection support device 100 has a processor 101 such as a CPU (Central Processing Unit) and a memory 102 that is a volatile storage device. The memory 102 is, for example, a volatile semiconductor memory such as a RAM (Random Access Memory). As shown in FIG. 1, the inspection support device 100 may have a nonvolatile storage device such as a hard disk drive (HDD) or solid state drive (SSD).
 点検支援装置100の各機能は、処理回路により実現される。処理回路は、専用のハードウェアであっても、メモリ102に格納されるプログラムを実行するプロセッサ101であってもよい。プロセッサ101は、処理装置、演算装置、マイクロプロセッサ、マイクロコンピュータ、及びDSP(Digital Signal Processor)のいずれであってもよい。 Each function of the inspection support device 100 is implemented by a processing circuit. The processing circuitry may be dedicated hardware or processor 101 executing a program stored in memory 102 . The processor 101 may be any of a processing device, an arithmetic device, a microprocessor, a microcomputer, and a DSP (Digital Signal Processor).
 処理回路が専用のハードウェアである場合、処理回路は、例えば、単一回路、複合回路、プログラム化したプロセッサ、並列プログラム化したプロセッサ、ASIC(Application Specific Integrated Circuit)、FPGA(Field-Programmable Gate Array)、又はこれらのうちのいずれかを組み合わせたものである。 If the processing circuit is dedicated hardware, the processing circuit may be, for example, a single circuit, a composite circuit, a programmed processor, a parallel programmed processor, an ASIC (Application Specific Integrated Circuit), an FPGA (Field-Programmable Gate Array) ), or a combination of any of these.
 処理回路がプロセッサ101である場合、点検支援プログラムは、ソフトウェア、ファームウェア、又はソフトウェアとファームウェアとの組み合わせにより実現される。ソフトウェア及びファームウェアは、プログラムとして記述され、メモリ102に格納される。プロセッサ101は、メモリ102に記憶された点検支援プログラムを読み出して実行することにより、図1に示される各部の機能を実現する。 When the processing circuit is the processor 101, the inspection support program is realized by software, firmware, or a combination of software and firmware. Software and firmware are written as programs and stored in memory 102 . The processor 101 reads out and executes the inspection support program stored in the memory 102, thereby implementing the functions of the units shown in FIG.
 なお、点検支援装置100は、一部を専用のハードウェアで実現し、一部をソフトウェア又はファームウェアで実現するようにしてもよい。このように、処理回路は、ハードウェア、ソフトウェア、ファームウェア、又はこれらのうちのいずれかの組み合わせによって、上述の各機能を実現することができる。 It should be noted that the inspection support device 100 may be partially realized by dedicated hardware and partially realized by software or firmware. As such, the processing circuitry may implement each of the functions described above in hardware, software, firmware, or any combination thereof.
 図5は、実施の形態1に係る点検支援装置100a及び携帯端末10aのハードウェア構成の他の例を示す図である。図5の例では、点検支援装置100aは、携帯端末10aと通信可能な装置である。携帯端末10aは、例えば、スマートフォン、タブレット型コンピュータ、又はPC、などである。携帯端末10aは、プロセッサ105と、メモリ106と、記憶部107と、通信部108と、カメラ110と、慣性センサ210と、表示装置180と、入力部190とを有している。点検支援装置100aは、プロセッサ101と、メモリ102と、記憶部103と、通信部104とを有している。 FIG. 5 is a diagram showing another example of the hardware configuration of the inspection support device 100a and the mobile terminal 10a according to the first embodiment. In the example of FIG. 5, the inspection support device 100a is a device capable of communicating with the mobile terminal 10a. The mobile terminal 10a is, for example, a smart phone, a tablet computer, or a PC. Portable terminal 10 a has processor 105 , memory 106 , storage unit 107 , communication unit 108 , camera 110 , inertial sensor 210 , display device 180 , and input unit 190 . The inspection support device 100 a has a processor 101 , a memory 102 , a storage section 103 and a communication section 104 .
 図5に示されるように、点検支援装置100aは、CPUなどのプロセッサ101とメモリ102とを有する。メモリ102は、例えば、RAMなどの、揮発性の半導体メモリである。図1に示されるように、点検支援装置100aは、HDD又はSSDなどの不揮発性の記憶装置を有してもよい。図5の例では、携帯端末10aと点検支援装置100aの組合が点検支援システムである。 As shown in FIG. 5, the inspection support device 100a has a processor 101 such as a CPU and a memory 102. Memory 102 is, for example, a volatile semiconductor memory such as RAM. As shown in FIG. 1, the inspection support device 100a may have a non-volatile storage device such as an HDD or SSD. In the example of FIG. 5, the combination of the mobile terminal 10a and the inspection support device 100a is the inspection support system.
 実施の形態1の処理は、端末位置姿勢推定スレッド、点検対象取得スレッド、及びAR表示スレッドの3つを並列して処理する。図6は、実施の形態1に係る点検支援装置100による端末位置姿勢推定スレッドにおける端末位置姿勢推定処理S11を示すフローチャートである。端末位置姿勢推定部120は、カメラ110から現在のカメラ画像と、センサから現在のセンサ情報を取得し、現在の携帯端末の位置姿勢を推定し、点検対象位置推定部150と表示制御部170に位置姿勢を通知する。 In the processing of the first embodiment, the terminal position/orientation estimation thread, the inspection target acquisition thread, and the AR display thread are processed in parallel. FIG. 6 is a flowchart showing the terminal position/orientation estimation processing S11 in the terminal position/orientation estimation thread by the inspection support apparatus 100 according to the first embodiment. The terminal position/orientation estimation unit 120 acquires the current camera image from the camera 110 and the current sensor information from the sensor, estimates the current position/orientation of the mobile terminal, and provides the inspection target position estimation unit 150 and the display control unit 170 Notify the position and orientation.
 図7は、実施の形態1に係る点検支援装置100による点検対象取得スレッドを示すフローチャートである。点検対象取得スレッドでは、点検対象検出部130は、カメラ110から現在のカメラ画像を取得し、カメラ画像中から0以上の点検対象領域を検出する(ステップS21)。 FIG. 7 is a flowchart showing an inspection target acquisition thread by the inspection support device 100 according to the first embodiment. In the inspection object acquisition thread, the inspection object detection unit 130 acquires the current camera image from the camera 110 and detects zero or more inspection object areas from the camera image (step S21).
 次に、点検対象追跡部140は、カメラ画像間の点検対象領域を時系列で追跡する(ステップS22)。 Next, the inspection target tracking unit 140 tracks the inspection target area between the camera images in time series (step S22).
 次に、それぞれの点検対象領域ごとに、以下のループ処理を行う(ステップS23)。
 先ず、点検対象位置推定部150は、現在の携帯端末10の位置姿勢と、追跡した点検対象領域とに基づいて、点検対象領域の位置(すなわち、3次元座標)を推定する(ステップS24)。次に、上記で推定した座標と、点検対象記憶部160が記憶しているそれぞれの点検対象の座標とを比較し、同じ座標であるか若しくは座標間の距離が一定値以内であった場合(ステップS25,S26)、当該点検対象は既に点検対象記憶部160に追加済みであると判断して、ループ処理の最初に戻り、他の場合は、当該点検対象を点検対象記憶部160に追加する(ステップS27)。
Next, the following loop processing is performed for each inspection target area (step S23).
First, the inspection target position estimation unit 150 estimates the position (ie, three-dimensional coordinates) of the inspection target region based on the current position and orientation of the mobile terminal 10 and the tracked inspection target region (step S24). Next, the coordinates estimated above are compared with the coordinates of each inspection object stored in the inspection object storage unit 160, and if the coordinates are the same or the distance between the coordinates is within a certain value ( Steps S25, S26), it is determined that the inspection object has already been added to the inspection object storage unit 160, the loop processing returns to the beginning, and in other cases, the inspection object is added to the inspection object storage unit 160. (Step S27).
 AR表示スレッドでは、表示制御部170は、カメラ110から現在のカメラ画像と、端末位置姿勢推定部120から現在の携帯端末の位置姿勢と、点検対象記憶部160から複数の点検対象を取得し、ARを画面に表示する。 In the AR display thread, the display control unit 170 acquires the current camera image from the camera 110, the current position and orientation of the mobile terminal from the terminal position/orientation estimation unit 120, and a plurality of inspection targets from the inspection target storage unit 160, Display the AR on the screen.
 以上に説明したように、実施の形態1に係る点検支援装置100及び携帯端末10を用いれば、点検対象にマーカを貼り付ける作業、3次元モデルの作成などを事前に行う必要がなく、また、事前に点検対象の位置情報を作成しなくても、ARで点検対象を表示することができる。点検対象の3次元座標を推定してARを表示するので、点検対象が検出されなくなった場合、又は、物陰に隠れた場合であっても、点検対象の位置を示すARを表示し続けることができる。 As described above, by using the inspection support device 100 and the mobile terminal 10 according to Embodiment 1, there is no need to perform operations such as attaching a marker to an inspection target and creating a three-dimensional model in advance. The object to be inspected can be displayed by AR without creating location information of the object to be inspected in advance. Since the AR is displayed by estimating the three-dimensional coordinates of the inspection target, even if the inspection target is no longer detected or is hidden behind something, the AR indicating the position of the inspection target can be continuously displayed. can.
実施の形態2.
 図8は、実施の形態2に係る点検支援装置200及び携帯端末20の構成を概略的に示す機能ブロック図である。図8において、図1に示される構成要素と同一又は対応する構成要素には、図1に示される符号と同じ符号が付されている。点検支援装置200は、文字認識部220及び認識対象取得部230を有する点、及び端末位置姿勢推定部120が慣性センサ210の出力を用いて位置及び姿勢を推定している点において、点検支援装置100と異なる。
Embodiment 2.
FIG. 8 is a functional block diagram schematically showing configurations of inspection support device 200 and mobile terminal 20 according to Embodiment 2. As shown in FIG. 8, the same or corresponding components as those shown in FIG. 1 are given the same reference numerals as those shown in FIG. Inspection support apparatus 200 is an inspection support apparatus in that it includes character recognition unit 220 and recognition target acquisition unit 230, and terminal position/orientation estimation unit 120 estimates the position and orientation using the output of inertial sensor 210. Different from 100.
 図9は、実施の形態2に係る携帯端末20のカメラ画像の表示例(文字列読取時)を示す図である。文字認識部220は、カメラ110によって取得されたカメラ画像から、点検対象であるバルブ821を識別する銘板821aの文字列である識別番号を認識する。認識処理は、ユーザが指示した際に実行される、又は、常時実行される。 FIG. 9 is a diagram showing a display example of a camera image of the mobile terminal 20 according to Embodiment 2 (at the time of reading a character string). The character recognition unit 220 recognizes, from the camera image acquired by the camera 110, the identification number, which is the character string of the nameplate 821a that identifies the valve 821 to be inspected. Recognition processing is executed when instructed by the user or is always executed.
 認識対象取得部230は、端末位置姿勢推定部120によって推定された携帯端末20の位置及び姿勢に基づいて、文字認識部220が認識した識別番号が付された点検対象に相当する、点検対象記憶部160が記憶する点検対象を取得する。すなわち、点検対象記憶部160に記憶されている複数の点検対象の3次元座標の中から、文字認識部220が認識した点検対象に相当する3次元座標を持つ点検対象を検索する。具体的には、認識対象取得部230は、例えば、携帯端末20の現在位置に最も近い3次元座標を持つ点検対象を点検対象記憶部160において検索し、今回の認識対象とする。或いは、認識対象取得部230は、携帯端末20の現在位置からカメラ110の視点方向に任意の距離(例えば、30cm)進んだ位置から最も近い3次元座標を持つ点検対象を検索し、今回の認識対象とする。また、認識対象取得部230は、点検対象記憶部160に記憶されている点検対象に「点検済み」の情報を付加して、未点検の点検対象と異なる表示としてもよい。このように、点検対象記憶部160は、実施の形態1で説明した情報に加え、それぞれの点検対象が点検済みか否かを表す情報を保持する。 Based on the position and orientation of the mobile terminal 20 estimated by the terminal position/orientation estimation unit 120, the recognition target acquisition unit 230 stores an inspection target corresponding to the inspection target with the identification number recognized by the character recognition unit 220. The inspection target stored by the unit 160 is acquired. That is, among the three-dimensional coordinates of a plurality of inspection objects stored in the inspection object storage unit 160, an inspection object having three-dimensional coordinates corresponding to the inspection object recognized by the character recognition unit 220 is searched. Specifically, for example, the recognition target acquisition unit 230 searches the inspection target storage unit 160 for an inspection target having three-dimensional coordinates closest to the current position of the mobile terminal 20, and sets it as the current recognition target. Alternatively, the recognition target acquiring unit 230 searches for the inspection target having the closest three-dimensional coordinates from a position that is an arbitrary distance (for example, 30 cm) in the viewing direction of the camera 110 from the current position of the mobile terminal 20, and performs the current recognition. set to target. In addition, the recognition target acquisition unit 230 may add the information “checked” to the inspection target stored in the inspection target storage unit 160 to display the inspection target differently from the uninspected inspection target. In this manner, the inspection target storage unit 160 holds information indicating whether or not each inspection target has been inspected in addition to the information described in the first embodiment.
 図10は、実施の形態2に係る携帯端末20の表示装置180の表示例を示す図である。表示制御部170は、認識対象取得部230によって認識された点検対象が点検済みか否かに応じて点検対象の表示方法を変える。表示は、例えば、ARを用いて行われる。例えば、点検済みの点検対象を囲うAR画像の枠の色と、点検前の点検対象を囲うAR画像の枠の色とを、異なる色にする。図10には、カメラ画像にAR画像を重ねて表示した例を示しているが、スマートグラスを用いることによって現実の風景に重ねてAR画像を表示することができる。また、プロジェクタを用いることによって、現実の設備(例えば、配管、バルブ)上にAR画像を投影してもよい。また、携帯端末20がカメラ画像を模擬したイラスト図形を形成するアプリを有する場合には、カメラ画像を模擬したイラスト上にAR画像を重畳表示してもよい。 FIG. 10 is a diagram showing a display example of the display device 180 of the mobile terminal 20 according to the second embodiment. The display control unit 170 changes the display method of the inspection object according to whether the inspection object recognized by the recognition object acquisition unit 230 has been inspected. The display is performed using AR, for example. For example, the color of the frame of the AR image that surrounds the inspection target that has already been inspected and the color of the frame of the AR image that surrounds the inspection target that has not yet been inspected are set to different colors. FIG. 10 shows an example in which an AR image is displayed superimposed on a camera image. By using smart glasses, an AR image can be displayed superimposed on a real landscape. Also, by using a projector, AR images may be projected onto real installations (eg, pipes, valves). Also, if the mobile terminal 20 has an application for forming an illustration figure simulating a camera image, the AR image may be superimposed on the illustration simulating the camera image.
 以上に説明したように、実施の形態2に係る点検支援装置200及び携帯端末20を用いれば、文字認識により点検対象の識別番号を認識し、当該点検対象を点検済みとして点検対象記憶部160が記憶することで、それぞれの点検対象が点検済みか否かをAR画像で区別することができる。また、文字認識により点検済みとすることで、識別番号の入力ミスの防止が可能である。 As described above, if the inspection support device 200 and the mobile terminal 20 according to the second embodiment are used, the identification number of the inspection object is recognized by character recognition, and the inspection object storage unit 160 determines that the inspection object has been inspected. By storing the AR image, it is possible to distinguish whether or not each inspection object has been inspected. In addition, it is possible to prevent mistakes in inputting the identification number by making the inspection completed by character recognition.
 上記以外に関し、実施の形態2は、実施の形態1と同じである。 Except for the above, the second embodiment is the same as the first embodiment.
実施の形態3.
 図11は、実施の形態3に係る点検支援装置300及び携帯端末30の構成を概略的に示す機能ブロック図である。図11において、図8に示される構成要素と同一又は対応する構成要素には、図8に示される符号と同じ符号が付されている。実施の形態3に係る点検支援装置300は、認識対象取得部230aの機能の点において、実施の形態2に係る点検支援装置200と異なる。
Embodiment 3.
FIG. 11 is a functional block diagram schematically showing configurations of inspection support device 300 and mobile terminal 30 according to the third embodiment. 11, the same or corresponding components as those shown in FIG. 8 are given the same reference numerals as those shown in FIG. Inspection support apparatus 300 according to Embodiment 3 differs from inspection support apparatus 200 according to Embodiment 2 in terms of the function of recognition target acquisition section 230a.
 点検対象位置推定部150は、携帯端末30の位置姿勢と複数枚のカメラ画像のうちの少なくとも2枚のカメラ画像における点検対象領域とに基づいて、点検対象領域の3次元座標系における座標を推定し、点検対象記憶部160に記憶させる。表示制御部170は、表示装置180に、表示部品(例えば、AR表示)及び認識された文字を表示させる。 The inspection target position estimating unit 150 estimates the coordinates of the inspection target region in the three-dimensional coordinate system based on the position and orientation of the mobile terminal 30 and the inspection target regions in at least two of the plurality of camera images. and stored in the inspection target storage unit 160 . The display control unit 170 causes the display device 180 to display display components (for example, AR display) and recognized characters.
 認識対象取得部230aは、点検対象記憶部160が記憶する点検対象を「点検済み」と書き換える際に、文字認識部220が認識した識別番号も記憶させる。 The recognition target acquisition unit 230a also stores the identification number recognized by the character recognition unit 220 when rewriting the inspection target stored in the inspection target storage unit 160 to "inspected".
 点検対象記憶部160は、それぞれの点検対象の識別番号も記憶する。未点検の場合は、識別番号不明とする。表示制御部170は、点検対象をARで表示する際に、当該点検対象の識別番号が判明している場合、識別番号もARで表示する。 The inspection target storage unit 160 also stores the identification number of each inspection target. If it has not been inspected, the identification number shall be unknown. When the inspection target is displayed in AR, the display control unit 170 also displays the identification number in AR if the identification number of the inspection target is known.
 図12は、実施の形態3に係る携帯端末30の表示装置180の表示例を示す図である。図12は、点検対象であるバルブ821の周辺(すなわち、近傍)に識別番号821cを表示している。 FIG. 12 is a diagram showing a display example of the display device 180 of the mobile terminal 30 according to the third embodiment. FIG. 12 displays an identification number 821c around (that is, in the vicinity of) the valve 821 to be inspected.
 以上に説明したように、実施の形態3に係る点検支援装置300及び携帯端末30によれば、認識済みの点検対象の近傍にAR画像によって識別番号を表示することで、点検済みの点検対象の判別が容易になる。 As described above, according to the inspection support device 300 and the mobile terminal 30 according to the third embodiment, by displaying the identification number in the vicinity of the recognized inspection object by the AR image, easier to distinguish.
 上記以外に関し、実施の形態3は、実施の形態1又は2と同じである。 Except for the above, Embodiment 3 is the same as Embodiment 1 or 2.
実施の形態4.
 図13は、実施の形態4に係る点検支援装置400及び携帯端末40の構成を概略的に示す機能ブロック図である。図13において、図11に示される構成要素と同一又は対応する構成要素には、図11に示される符号と同じ符号が付されている。実施の形態4に係る点検支援装置400は、端末位置姿勢記憶部410及び端末位置姿勢復帰部420を有する点において、実施の形態3に係る点検支援装置300と異なる。
Embodiment 4.
FIG. 13 is a functional block diagram schematically showing configurations of inspection support device 400 and portable terminal 40 according to the fourth embodiment. 13, the same or corresponding components as those shown in FIG. 11 are given the same reference numerals as those shown in FIG. Inspection support apparatus 400 according to Embodiment 4 differs from inspection support apparatus 300 according to Embodiment 3 in that it includes terminal position/posture storage section 410 and terminal position/posture recovery section 420 .
 実施の形態4では、端末位置姿勢推定部120は、カメラ画像と携帯端末40の位置姿勢とを関連付けて端末位置姿勢記憶部410に記憶させる。端末位置姿勢復帰部420は、端末位置姿勢推定部120に記憶されているカメラ画像のうちの、現在のカメラ画像に最も近いカメラ画像に紐付いた位置姿勢を、現在の位置姿勢として端末位置姿勢推定部120に提供する。 In Embodiment 4, the terminal position/orientation estimation unit 120 associates the camera image with the position/orientation of the mobile terminal 40 and stores them in the terminal position/orientation storage unit 410 . The terminal position/orientation restoring unit 420 estimates the terminal position/orientation using the position/orientation associated with the camera image closest to the current camera image among the camera images stored in the terminal position/orientation estimation unit 120 as the current position/orientation. 120.
 端末位置姿勢記憶部410は、定期的(例えば、10秒ごと)に、現在のカメラ画像と、端末位置姿勢推定部120が推定した現在の携帯端末の位置姿勢との関係性を記憶する。この関係性は、複数個保持される。カメラ画像は、画像の生データではなく、後述の画像同士のマッチングが可能な形式であれば、処理・加工されたデータでもよい。データは、局所特徴量であるSIFT(Scale-invariant feature transform)、SURF(Speed-Upped Robust Feature)、ORB(Oriented FAST and Rotated BRIEF)、KAZEのような形式でもよい。 The terminal position/orientation storage unit 410 periodically (for example, every 10 seconds) stores the relationship between the current camera image and the current position/orientation of the mobile terminal estimated by the terminal position/orientation estimation unit 120 . A plurality of such relationships are held. The camera image may not be raw data of the image, but may be processed/processed data as long as it is in a format that enables matching between images, which will be described later. The data may be in formats such as SIFT (Scale-invariant feature transform), SURF (Speed-Upped Robust Feature), ORB (Oriented FAST and Rotated BRIEF), and KAZE, which are local features.
 端末位置姿勢復帰部420は、点検支援装置の起動時に、現在のカメラ画像と、端末位置姿勢記憶部410が記憶するそれぞれのカメラ画像とをマッチングし、最も現在のカメラ画像に近い、端末位置姿勢記憶部410が記憶するカメラ画像を特定する。特定に成功したら、当該カメラ画像に紐付いた携帯端末の位置姿勢を、現在の携帯端末の位置姿勢として、端末位置姿勢推定部120を初期化する。 The terminal position/orientation recovery unit 420 matches the current camera image with each camera image stored in the terminal position/orientation storage unit 410 when the inspection support device is activated, and returns the terminal position/orientation closest to the current camera image. A camera image stored in the storage unit 410 is specified. When the identification is successful, the terminal position/orientation estimating unit 120 is initialized with the position/orientation of the mobile terminal linked to the camera image as the current position/orientation of the mobile terminal.
 以上に説明したように、実施の形態4に係る点検支援装置400及び携帯端末40を用いれば、定期的にカメラ画像と携帯端末40の位置姿勢を記憶しておき、携帯端末40の起動時に現在のカメラ画像と過去に記憶したカメラ画像とをマッチングすることで、前回の3次元座標系を引き継いで点検作業を再開することができる。したがって、再度点検対象の検出及び位置推定などを行う必要なく、直ぐにAR表示を再開することができる。 As described above, by using the inspection support device 400 and the mobile terminal 40 according to the fourth embodiment, the camera image and the position and orientation of the mobile terminal 40 are periodically stored, and when the mobile terminal 40 is activated, the current By matching the previous camera image with the camera image stored in the past, the previous three-dimensional coordinate system can be inherited and the inspection work can be restarted. Therefore, the AR display can be resumed immediately without the need to detect and estimate the position of the object to be inspected again.
 上記以外に関し、実施の形態4は、実施の形態1から3のいずれかと同じである。 Except for the above, Embodiment 4 is the same as any of Embodiments 1 to 3.
実施の形態5.
 図14は、実施の形態5に係る点検支援装置500及び携帯端末50の構成を概略的に示す機能ブロック図である。図14において、図13に示される構成要素と同一又は対応する構成要素には、図13に示される符号と同じ符号が付されている。実施の形態5に係る点検支援装置500は、点検状態初期化部510を有する点において、実施の形態4に係る点検支援装置400と異なる。
Embodiment 5.
FIG. 14 is a functional block diagram schematically showing configurations of inspection support device 500 and mobile terminal 50 according to Embodiment 5. As shown in FIG. 14, the same or corresponding components as those shown in FIG. 13 are given the same reference numerals as those shown in FIG. Inspection support device 500 according to Embodiment 5 differs from inspection support device 400 according to Embodiment 4 in that inspection state initialization section 510 is provided.
 点検状態初期化部510は、点検支援装置500の起動時又は入力部(図4の190)を経由してユーザから指定された際に、点検対象記憶部160が保持する一部又はすべての点検対象の点検状態の情報を、未点検とする。 The inspection state initialization unit 510 initializes some or all of the inspections held by the inspection target storage unit 160 when the inspection support device 500 is activated or when specified by the user via the input unit (190 in FIG. 4). Information on the inspection status of the target is set to uninspected.
 以上に説明したように、実施の形態5に係る点検支援装置500及び携帯端末50を用いれば、新たな点検を開始することができる。つまり、点検対象の座標を表すために用いられる3次元座標は、既に一度求められているので、再度、点検対象の検出及び座標の推定を行う必要がない場合があり、すべての点検対象がAR表示された状態で点検を開始することができる。また、識別番号も判明しているため、識別番号に基づいて未点検の点検対象を探すことも可能である。 As described above, a new inspection can be started using the inspection support device 500 and the mobile terminal 50 according to the fifth embodiment. In other words, since the three-dimensional coordinates used to represent the coordinates of the inspection target have already been obtained once, it may not be necessary to detect the inspection target and estimate the coordinates again. Inspection can be started in the displayed state. In addition, since the identification number is also known, it is possible to search for an uninspected object to be inspected based on the identification number.
 上記以外に関し、実施の形態5は、実施の形態1から4のいずれかと同じである。 Except for the above, Embodiment 5 is the same as any of Embodiments 1 to 4.
実施の形態6.
 図15は、実施の形態6に係る点検支援装置600及び携帯端末60の構成を概略的に示す機能ブロック図である。図15において、図14に示される構成要素と同一又は対応する構成要素には、図14に示される符号と同じ符号が付されている。実施の形態6に係る点検支援装置600は、点検手順記憶部610及び点検手順管理部620を有する点において、実施の形態5に係る点検支援装置500と異なる。
Embodiment 6.
FIG. 15 is a functional block diagram schematically showing configurations of inspection support device 600 and mobile terminal 60 according to Embodiment 6. As shown in FIG. 15, the same or corresponding components as those shown in FIG. 14 are given the same reference numerals as those shown in FIG. The inspection support device 600 according to the sixth embodiment differs from the inspection support device 500 according to the fifth embodiment in that it has an inspection procedure storage unit 610 and an inspection procedure management unit 620 .
 点検手順記憶部610は、予め作成した点検手順を示す情報を記憶する。点検手順を示す情報は、複数の点検対象の点検の順番を示す情報であり、具体的には、整列された識別番号のリストである。 The inspection procedure storage unit 610 stores information indicating an inspection procedure created in advance. The information indicating the inspection procedure is information indicating the order of inspection of a plurality of inspection targets, and is specifically a list of arranged identification numbers.
 点検手順管理部620は、現在の点検手順の状態を管理する。文字認識部220によって識別番号を読み取り、かつ、その識別番号が現在の点検手順の識別番号であった場合、点検手順を1つ進める。現在の点検手順の情報は、表示制御部670に提供される。つまり、点検手順管理部620は、点検手順記憶部610に記憶されている点検対象の点検の順番と認識された文字とに基づいて、表示制御部670に情報を提供して、次の点検対象を示すAR表示を表示装置180に実行させる。 The inspection procedure management unit 620 manages the current inspection procedure status. If the identification number is read by the character recognition unit 220 and the identification number is the identification number of the current inspection procedure, the inspection procedure is advanced by one. Information on the current inspection procedure is provided to the display control unit 670 . That is, the inspection procedure management unit 620 provides information to the display control unit 670 based on the order of inspection of the inspection objects stored in the inspection procedure storage unit 610 and the recognized characters, and displays the next inspection object. is displayed on the display device 180 .
 表示制御部670は、点検対象をARで表示する際に、当該点検対象の識別番号が、点検手順管理部620が管理している現在点検中の点検対象の識別番号であった場合、当該点検対象のARの表示が他の点検対象と比べて目立つように表示する。表示制御部670は、例えば、点検手順管理部620が管理している現在点検中の点検対象の識別番号であった場合、当該点検対象についてのAR表示が他の点検対象と比べて目立つように、AR表示の色を変更させる、又は、点検対象を示す表示部品を点滅させる。 When the inspection target is displayed by AR, if the identification number of the inspection target is the identification number of the current inspection target managed by the inspection procedure management unit 620, the display control unit 670 displays the inspection target. Display the target AR so that it stands out compared to other inspection targets. For example, when it is the identification number of an inspection object currently being inspected managed by the inspection procedure management unit 620, the display control unit 670 makes the AR display for the inspection object stand out compared to other inspection objects. , change the color of the AR display, or blink the display part indicating the inspection target.
 図16は、実施の形態6に係る携帯端末60の表示装置180の表示例を示す図である。図16は、次の点検対象が、バルブ823であることを示している。次の点検対象の表示方法は、図16のものに限定されない。 FIG. 16 is a diagram showing a display example of the display device 180 of the mobile terminal 60 according to the sixth embodiment. FIG. 16 shows that the valve 823 is to be inspected next. The display method of the next inspection target is not limited to that of FIG.
 以上に説明したように、実施の形態6に係る点検支援装置600及び携帯端末60を用いれば、点検対象の位置をARで表示することができるため、直感的に手順通りに点検することができる。この場合には、予め点検手順のデータを作成する必要があるが、点検の順序を示すデータであるから、作成の労力は少なく、また作成コストは低い。 As described above, by using the inspection support device 600 and the mobile terminal 60 according to Embodiment 6, it is possible to display the position of the inspection target in AR, so that the inspection can be performed intuitively according to the procedure. . In this case, it is necessary to create inspection procedure data in advance, but since the data indicates the order of inspection, the creation effort is small and the creation cost is low.
 上記以外に関し、実施の形態6は、実施の形態1から5のいずれかと同じである。 Except for the above, Embodiment 6 is the same as any of Embodiments 1 to 5.
実施の形態7.
 図17は、実施の形態7に係る点検支援装置700及び携帯端末70の構成を概略的に示す機能ブロック図である。図17において、図14に示される構成要素と同一又は対応する構成要素には、図14に示される符号と同じ符号が付されている。実施の形態7に係る点検支援装置700は、識別番号確認部710を有する点において、実施の形態5に係る点検支援装置500と異なる。
Embodiment 7.
FIG. 17 is a functional block diagram schematically showing configurations of inspection support device 700 and mobile terminal 70 according to Embodiment 7. As shown in FIG. 17, the same or corresponding components as those shown in FIG. 14 are given the same reference numerals as those shown in FIG. Inspection support device 700 according to Embodiment 7 differs from inspection support device 500 according to Embodiment 5 in that it includes an identification number confirmation unit 710 .
 識別番号確認部710は、文字認識部220によって今回認識された識別番号が、点検対象記憶部160に既に記憶されている同じ点検対象についての識別番号(すなわち、過去に認識された識別番号)と一致しない場合、警告を表示装置180の画面に表示させる識別番号確認部を更に有している。例えば、文字認識部220及び認識対象取得部230が、点検対象記憶部160に記憶されている点検対象の点検状態情報を「未点検」から「点検済み」に変更する際に、点検対象記憶部160によって記憶されている当該点検対象が既に識別番号を記憶しており、かつ、当該識別番号と、今回認識された識別番号とが異なっていた場合、識別番号確認部710は、警告を表示装置180の画面に表示させる。 Identification number confirmation unit 710 confirms that the identification number recognized this time by character recognition unit 220 is the identification number for the same inspection object already stored in inspection object storage unit 160 (that is, the identification number recognized in the past). It further has an identification number confirming unit for displaying a warning on the screen of the display device 180 when the numbers do not match. For example, when the character recognition unit 220 and the recognition target acquisition unit 230 change the inspection status information of the inspection target stored in the inspection target storage unit 160 from "unchecked" to "checked", the inspection target storage unit 160 has already stored an identification number, and the identification number is different from the identification number recognized this time, the identification number confirmation unit 710 displays a warning on the display device. 180 screen.
 図18は、実施の形態7に係る携帯端末70の表示装置180の表示例を示す図である。図18に示されるように、識別番号確認部710は、今回認識された識別番号が、点検対象記憶部160に既に記憶されている同じ点検対象についての識別番号と一致しないことを示す、警告を表示装置180の画面に表示させる。警告は、図18に示されるものに限定されず、他の文章、映像、音声などであってもよい。 FIG. 18 is a diagram showing a display example of the display device 180 of the mobile terminal 70 according to the seventh embodiment. As shown in FIG. 18 , identification number confirmation unit 710 issues a warning indicating that the identification number recognized this time does not match the identification number for the same inspection object already stored in inspection object storage unit 160 . It is displayed on the screen of the display device 180 . The warning is not limited to that shown in FIG. 18, and may be other text, video, audio, and the like.
 以上に説明したように、実施の形態7に係る点検支援装置700及び携帯端末70を用いれば、同一の点検対象に対して、過去に認識した識別番号と、今回認識した識別番号とが異なっているという異常な事態をユーザに知らせることができる。ここで、異常な事態は、例えば、現場の設備の点検対象の配置が変更されている、点検対象の銘板が変更されている、などである。 As described above, if the inspection support device 700 and the mobile terminal 70 according to the seventh embodiment are used, the identification number recognized in the past and the identification number recognized this time are different for the same inspection object. It is possible to inform the user of an abnormal situation that there is Here, the abnormal situation is, for example, a change in the arrangement of inspection targets in the field equipment, a change in the inspection target nameplate, or the like.
 上記以外に関し、実施の形態7は、実施の形態1から6のいずれかと同じである。 Except for the above, Embodiment 7 is the same as any of Embodiments 1 to 6.
 10、20、30、40、50、60、70 携帯端末、 100、200、300、400、500、600、700 点検支援装置、 10a 携帯端末、 100a 点検支援装置、 110 カメラ、 120 端末位置姿勢推定部、 130 点検対象検出部、 140 点検対象追跡部、 150 点検対象位置推定部、 160 点検対象記憶部、 170、670 表示制御部、 180 表示装置、 210 慣性センサ、 220 文字認識部、 230、230a 認識対象取得部、 410 端末位置姿勢記憶部、 420 端末位置姿勢復帰部、 510 点検状態初期化部、 610 点検手順記憶部、 620 点検手順管理部、 710 識別番号確認部、 800 ユーザ、 810 配管、 821~825 バルブ(点検対象)、 821a~825a 銘板、 821b~825b 点検対象領域、 821c 識別番号、 830 タンク。 10, 20, 30, 40, 50, 60, 70 mobile terminal, 100, 200, 300, 400, 500, 600, 700 inspection support device, 10a mobile terminal, 100a inspection support device, 110 camera, 120 terminal position and orientation estimation 130 Inspection object detection unit 140 Inspection object tracking unit 150 Inspection object position estimation unit 160 Inspection object storage unit 170, 670 Display control unit 180 Display device 210 Inertial sensor 220 Character recognition unit 230, 230a Recognition target acquisition unit 410 Terminal position/orientation storage unit 420 Terminal position/orientation return unit 510 Inspection state initialization unit 610 Inspection procedure storage unit 620 Inspection procedure management unit 710 Identification number confirmation unit 800 User 810 Piping, 821 to 825 valves (to be inspected), 821a to 825a nameplates, 821b to 825b inspection target areas, 821c identification numbers, 830 tanks.

Claims (15)

  1.  携帯端末に備えられたカメラにより撮影されたカメラ画像と前記携帯端末に備えられた慣性センサにより得られた検出信号とのうちの少なくとも一方に基づいて、3次元座標系における前記携帯端末の位置姿勢を推定する端末位置姿勢推定部と、
     前記カメラにより撮影されたカメラ画像から、点検対象を検出する点検対象検出部と、
     前記カメラ画像において検出された前記点検対象を示す点検対象領域を撮影時刻順に時系列で追跡して、時系列に並ぶ複数枚のカメラ画像において同じ点検対象領域を特定する点検対象追跡部と、
     前記携帯端末の前記位置姿勢と前記複数枚のカメラ画像のうちの少なくとも2枚のカメラ画像における前記点検対象領域とに基づいて、前記点検対象領域の前記3次元座標系における座標を推定する点検対象位置推定部と、
     前記点検対象領域を視認可能にする表示部品を表示装置に表示させる表示制御部と、
     を有することを特徴とする点検支援装置。
    The position and orientation of the mobile terminal in a three-dimensional coordinate system based on at least one of a camera image captured by a camera provided in the mobile terminal and a detection signal obtained by an inertial sensor provided in the mobile terminal. a terminal position/orientation estimation unit for estimating
    an inspection target detection unit that detects an inspection target from a camera image captured by the camera;
    an inspection target tracking unit that tracks the inspection target area indicating the inspection target detected in the camera image in chronological order of photographing time and identifies the same inspection target area in a plurality of camera images arranged in time series;
    An inspection target for estimating coordinates of the inspection target area in the three-dimensional coordinate system based on the position and orientation of the mobile terminal and the inspection target area in at least two camera images out of the plurality of camera images. a position estimator;
    a display control unit that causes a display device to display a display component that makes the inspection target area visible;
    An inspection support device comprising:
  2.  前記点検対象位置推定部は、前記携帯端末の前記位置姿勢と前記点検対象領域の前記座標とを関連付けて点検対象記憶部に記憶させる
     ことを特徴とする請求項1に記載の点検支援装置。
    2. The inspection support apparatus according to claim 1, wherein the inspection target position estimating unit associates the position and orientation of the mobile terminal with the coordinates of the inspection target area and stores them in an inspection target storage unit.
  3.  前記表示制御部は、前記表示装置に、点検済の前記点検対象と未点検の前記点検対象とを互いに異なる表示部品で表示させる
     ことを特徴とする請求項1又は2に記載の点検支援装置。
    The inspection support device according to claim 1 or 2, wherein the display control unit causes the display device to display the inspection object that has been inspected and the inspection object that has not been inspected using different display parts.
  4.  前記点検対象位置推定部は、前記携帯端末の前記位置姿勢と前記複数枚のカメラ画像のうちの少なくとも2枚のカメラ画像における点検対象領域とに基づいて、前記点検対象領域の3次元座標系における前記座標を推定する
     ことを特徴とする請求項1に記載の点検支援装置。
    The inspection target position estimating unit calculates, based on the position and orientation of the mobile terminal and an inspection target region in at least two camera images among the plurality of camera images, a three-dimensional coordinate system of the inspection target region. The inspection support device according to claim 1, wherein the coordinates are estimated.
  5.  前記点検対象に備えられた銘板の文字を認識する文字認識部と、
     前記点検対象と前記認識された文字とを関連付ける認識対象取得部と、
     を更に有し、
     前記表示制御部は、前記表示装置に、前記表示部品及び前記認識された文字を表示させる
     ことを特徴とする請求項4に記載の点検支援装置。
    a character recognition unit that recognizes characters on a nameplate provided for the inspection target;
    a recognition target acquisition unit that associates the inspection target with the recognized character;
    further having
    The inspection support device according to claim 4, wherein the display control unit causes the display device to display the display component and the recognized characters.
  6.  端末位置姿勢復帰部を更に有し、
     前記端末位置姿勢推定部は、前記カメラ画像と前記携帯端末の前記位置姿勢とを関連付けて端末位置姿勢記憶部に記憶させ、
     前記端末位置姿勢復帰部は、前記端末位置姿勢推定部に記憶されている前記カメラ画像のうちの、現在のカメラ画像に最も近いカメラ画像に紐付いた前記位置姿勢を、現在の位置姿勢として前記端末位置姿勢推定部に提供する
     ことを特徴とする請求項1から5のいずれか1項に記載の点検支援装置。
    further comprising a terminal position/orientation restoring unit;
    The terminal position/orientation estimation unit stores the camera image and the position/orientation of the mobile terminal in a terminal position/orientation storage unit in association with each other;
    The terminal position/orientation restoring unit sets the position/orientation associated with the camera image closest to the current camera image among the camera images stored in the terminal position/orientation estimation unit as the current position/orientation of the terminal. The inspection support device according to any one of claims 1 to 5, wherein the information is provided to a position/orientation estimation unit.
  7.  前記点検対象記憶部に記憶されている前記点検対象の点検状態を、未点検に初期化する点検状態初期化部を更に有する
     ことを特徴とする請求項2に記載の点検支援装置。
    3. The inspection support device according to claim 2, further comprising an inspection state initialization unit that initializes the inspection state of the inspection object stored in the inspection object storage unit to uninspected.
  8.  前記点検対象に備えられた銘板の文字を認識する文字認識部と、
     点検手順記憶部に記憶されている前記点検対象の点検の順番と前記認識された文字とに基づいて、次の点検対象を示す表示を前記表示装置に実行させる点検手順管理部と、
     を更に有することを特徴とする請求項1から3のいずれか1項に記載の点検支援装置。
    a character recognition unit that recognizes characters on a nameplate provided for the inspection target;
    an inspection procedure management unit that causes the display device to display the next inspection object based on the order of inspection of the inspection object and the recognized characters stored in the inspection procedure storage unit;
    The inspection support device according to any one of claims 1 to 3, further comprising:
  9.  前記文字認識部によって今回認識された識別番号が、前記点検対象記憶部に既に記憶されている同じ点検対象についての識別番号と一致しない場合、警告を前記表示装置の画面に表示させる識別番号確認部を更に有する
     ことを特徴とする請求項4又は5に記載の点検支援装置。
    Identification number confirmation unit for displaying a warning on the screen of the display device when the identification number recognized this time by the character recognition unit does not match the identification number for the same inspection object already stored in the inspection object storage unit The inspection support device according to claim 4 or 5, further comprising:
  10.  前記表示装置は、画像を表示する画面を有し、
     前記表示制御部は、前記画面に、前記カメラ画像と前記カメラ画像に重なる前記表示部品とを表示させる
     ことを特徴とする請求項1から9のいずれか1項に記載の点検支援装置。
    The display device has a screen for displaying an image,
    The inspection support device according to any one of claims 1 to 9, wherein the display control unit causes the screen to display the camera image and the display component overlapping the camera image.
  11.  前記表示装置は、画像を表示する透明又は半透明の画面を有し、
     前記表示制御部は、前記画面に、前記画面を介して視認される現実の風景に重なる前記表示部品を表示させる
     ことを特徴とする請求項1から9のいずれか1項に記載の点検支援装置。
    The display device has a transparent or translucent screen for displaying an image,
    10. The inspection support device according to any one of claims 1 to 9, wherein the display control unit causes the screen to display the display component that overlaps a real landscape viewed through the screen. .
  12.  前記表示装置は、プロジェクタであり、
     前記表示制御部は、前記プロジェクタにより現実の設備上に前記表示部品を投影させる
     ことを特徴とする請求項1から9のいずれか1項に記載の点検支援装置。
    The display device is a projector,
    The inspection support device according to any one of claims 1 to 9, wherein the display control unit causes the projector to project the display component onto actual equipment.
  13.  請求項1から12のいずれか1項に記載の点検支援装置と、
     前記カメラと、
     前記慣性センサと、
     前記表示装置と
     を有することを特徴とする点検支援システム。
    The inspection support device according to any one of claims 1 to 12;
    the camera;
    the inertial sensor;
    An inspection support system comprising: the display device;
  14.  点検支援装置によって実行される点検支援方法であって、
     携帯端末に備えられたカメラにより撮影されたカメラ画像と前記携帯端末に備えられた慣性センサにより得られた検出信号とのうちの少なくとも一方に基づいて、3次元座標系における前記携帯端末の位置姿勢を推定するステップと、
     前記カメラにより撮影されたカメラ画像から、点検対象を検出するステップと、
     前記カメラ画像において検出された前記点検対象を示す点検対象領域を撮影時刻順に時系列で追跡して、時系列に並ぶ複数枚のカメラ画像において同じ点検対象領域を特定するステップと、
     前記携帯端末の前記位置姿勢と前記複数枚のカメラ画像のうちの少なくとも2枚のカメラ画像における前記点検対象領域とに基づいて、前記点検対象領域の前記3次元座標系における座標を推定するステップと、
     前記点検対象領域を視認可能にする表示部品を表示装置に表示させるステップと、
     を有することを特徴とする点検支援方法。
    An inspection support method executed by an inspection support device,
    The position and orientation of the mobile terminal in a three-dimensional coordinate system based on at least one of a camera image captured by a camera provided in the mobile terminal and a detection signal obtained by an inertial sensor provided in the mobile terminal. estimating the
    a step of detecting an inspection target from a camera image taken by the camera;
    a step of tracking the inspection target area indicating the inspection target detected in the camera image in chronological order of photographing time, and identifying the same inspection target area in a plurality of camera images arranged in time series;
    estimating coordinates of the inspection target area in the three-dimensional coordinate system based on the position and orientation of the mobile terminal and the inspection target area in at least two camera images out of the plurality of camera images; ,
    a step of displaying on a display device a display component that makes the inspection target area visible;
    An inspection support method characterized by having
  15.  携帯端末に備えられたカメラにより撮影されたカメラ画像と前記携帯端末に備えられた慣性センサにより得られた検出信号とのうちの少なくとも一方に基づいて、3次元座標系における前記携帯端末の位置姿勢を推定するステップと、
     前記カメラにより撮影されたカメラ画像から、点検対象を検出するステップと、
     前記カメラ画像において検出された前記点検対象を示す点検対象領域を撮影時刻順に時系列で追跡して、時系列に並ぶ複数枚のカメラ画像において同じ点検対象領域を特定するステップと、
     前記携帯端末の前記位置姿勢と前記複数枚のカメラ画像のうちの少なくとも2枚のカメラ画像における前記点検対象領域とに基づいて、前記点検対象領域の前記3次元座標系における座標を推定するステップと、
     前記点検対象領域を視認可能にする表示部品を表示装置に表示させるステップと、
     をコンピュータに実行させることを特徴とする点検支援プログラム。
    The position and orientation of the mobile terminal in a three-dimensional coordinate system based on at least one of a camera image captured by a camera provided in the mobile terminal and a detection signal obtained by an inertial sensor provided in the mobile terminal. estimating the
    a step of detecting an inspection target from a camera image taken by the camera;
    a step of tracking the inspection target area indicating the inspection target detected in the camera image in chronological order of photographing time, and identifying the same inspection target area in a plurality of camera images arranged in time series;
    estimating coordinates of the inspection target area in the three-dimensional coordinate system based on the position and orientation of the mobile terminal and the inspection target area in at least two camera images out of the plurality of camera images; ,
    a step of displaying on a display device a display component that makes the inspection target area visible;
    An inspection support program characterized by causing a computer to execute
PCT/JP2021/015597 2021-04-15 2021-04-15 Inspection assistance device, inspection assistance system, method for assisting inspection, and inspection assistance program WO2022219780A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP2023514277A JP7361992B2 (en) 2021-04-15 2021-04-15 Inspection support device, inspection support system, inspection support method, and inspection support program
PCT/JP2021/015597 WO2022219780A1 (en) 2021-04-15 2021-04-15 Inspection assistance device, inspection assistance system, method for assisting inspection, and inspection assistance program

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2021/015597 WO2022219780A1 (en) 2021-04-15 2021-04-15 Inspection assistance device, inspection assistance system, method for assisting inspection, and inspection assistance program

Publications (1)

Publication Number Publication Date
WO2022219780A1 true WO2022219780A1 (en) 2022-10-20

Family

ID=83640249

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/015597 WO2022219780A1 (en) 2021-04-15 2021-04-15 Inspection assistance device, inspection assistance system, method for assisting inspection, and inspection assistance program

Country Status (2)

Country Link
JP (1) JP7361992B2 (en)
WO (1) WO2022219780A1 (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004173108A (en) * 2002-11-21 2004-06-17 Central Res Inst Of Electric Power Ind Database generation method, check support system, and program for check support of sensor information
JP2008261803A (en) * 2007-04-13 2008-10-30 Chugoku Electric Power Co Inc:The Valve inspection support system
WO2013030929A1 (en) * 2011-08-29 2013-03-07 株式会社日立製作所 Monitoring device, monitoring system and monitoring method
JP2015170302A (en) * 2014-03-10 2015-09-28 株式会社建設環境研究所 Sluiceway and sluice pipe inspection support system, and sluiceway and sluice pipe inspection support method
JP2017163445A (en) * 2016-03-11 2017-09-14 古河電気工業株式会社 Inspection support system, inspection support method and program
JP2018190304A (en) * 2017-05-10 2018-11-29 株式会社日立製作所 Inspection support device, inspection support method, and inspection support program
WO2020144848A1 (en) * 2019-01-11 2020-07-16 三菱電機株式会社 Authoring device, authoring method, and authoring program
WO2020234912A1 (en) * 2019-05-17 2020-11-26 三菱電機株式会社 Mobile device, position display method, and position display program

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004173108A (en) * 2002-11-21 2004-06-17 Central Res Inst Of Electric Power Ind Database generation method, check support system, and program for check support of sensor information
JP2008261803A (en) * 2007-04-13 2008-10-30 Chugoku Electric Power Co Inc:The Valve inspection support system
WO2013030929A1 (en) * 2011-08-29 2013-03-07 株式会社日立製作所 Monitoring device, monitoring system and monitoring method
JP2015170302A (en) * 2014-03-10 2015-09-28 株式会社建設環境研究所 Sluiceway and sluice pipe inspection support system, and sluiceway and sluice pipe inspection support method
JP2017163445A (en) * 2016-03-11 2017-09-14 古河電気工業株式会社 Inspection support system, inspection support method and program
JP2018190304A (en) * 2017-05-10 2018-11-29 株式会社日立製作所 Inspection support device, inspection support method, and inspection support program
WO2020144848A1 (en) * 2019-01-11 2020-07-16 三菱電機株式会社 Authoring device, authoring method, and authoring program
WO2020234912A1 (en) * 2019-05-17 2020-11-26 三菱電機株式会社 Mobile device, position display method, and position display program

Also Published As

Publication number Publication date
JPWO2022219780A1 (en) 2022-10-20
JP7361992B2 (en) 2023-10-16

Similar Documents

Publication Publication Date Title
US11393173B2 (en) Mobile augmented reality system
US11079841B2 (en) Enabling augmented reality using eye gaze tracking
US9710970B2 (en) Method and apparatus for providing contents including augmented reality information
CN106471548B (en) Use the method and apparatus of the acceleration template matches of peripheral information
US11842514B1 (en) Determining a pose of an object from rgb-d images
CN107710280B (en) Object visualization method
CN104284146A (en) Tracking assistance device, tracking assistance system and tracking assistance method
US9785836B2 (en) Dataset creation for tracking targets with dynamically changing portions
US9058660B2 (en) Feature searching based on feature quality information
US11568551B2 (en) Method and system for obtaining pair-wise epipolar constraints and solving for panorama pose on a mobile device
CN110781823B (en) Screen recording detection method and device, readable medium and electronic equipment
US20120293550A1 (en) Localization device and localization method with the assistance of augmented reality
US20170085656A1 (en) Automatic absolute orientation and position
US9076062B2 (en) Feature searching along a path of increasing similarity
Santos et al. Hybrid approach using sensors, GPS and vision based tracking to improve the registration in mobile augmented reality applications
WO2022219780A1 (en) Inspection assistance device, inspection assistance system, method for assisting inspection, and inspection assistance program
CN103631962A (en) Display method and equipment for image label
US10108882B1 (en) Method to post and access information onto a map through pictures
US20140140573A1 (en) Pose Tracking through Analysis of an Image Pyramid
US20210278242A1 (en) Wearable terminal display system, wearable terminal display method and program
CN117671004A (en) Wearing equipment posture estimation method, electronic equipment, storage medium and wearing system
JP2024070133A (en) Target recognition method, program and device
JP2019057014A (en) Subject specification information processing device, and control method and control program thereof
EP3279850A1 (en) System and method for interactive shopping

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21936976

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2023514277

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21936976

Country of ref document: EP

Kind code of ref document: A1