WO2024042644A1 - Video processing device, video processing method, and video processing program - Google Patents

Video processing device, video processing method, and video processing program Download PDF

Info

Publication number
WO2024042644A1
WO2024042644A1 PCT/JP2022/031904 JP2022031904W WO2024042644A1 WO 2024042644 A1 WO2024042644 A1 WO 2024042644A1 JP 2022031904 W JP2022031904 W JP 2022031904W WO 2024042644 A1 WO2024042644 A1 WO 2024042644A1
Authority
WO
WIPO (PCT)
Prior art keywords
sight
line
video processing
unit
image
Prior art date
Application number
PCT/JP2022/031904
Other languages
French (fr)
Japanese (ja)
Inventor
誠 武藤
Original Assignee
日本電信電話株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日本電信電話株式会社 filed Critical 日本電信電話株式会社
Priority to PCT/JP2022/031904 priority Critical patent/WO2024042644A1/en
Publication of WO2024042644A1 publication Critical patent/WO2024042644A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics

Definitions

  • the present invention relates to a video processing device, a video processing method, and a video processing program.
  • a user using an augmented reality (AR) system can view the real space of the real world through a mobile terminal or an AR device.
  • content such as navigation information or 3D data (hereinafter referred to as AR content) is presented as additional information in the real space. That is, a user using an AR system can see AR content superimposed on the real world and use information about this content.
  • the image displayed by the AR device will include a portion of the camera image showing the environment (scenery in front of the bicycle) and a portion showing the moving object. (a part of the bicycle body).
  • This invention has been made in view of the above-mentioned circumstances, and its purpose is to prevent mobile terminals or
  • the purpose of an AR device is to provide a technology that can display AR content at a precise location.
  • one aspect of the present invention is an image processing device worn by a user, which acquires an image captured by a camera and including a moving object on which the user is riding and an environment. a moving object detecting section that detects the moving object from the photographed image; and a moving object detecting section that detects the moving object from the photographed image, and complements an image of the environment in the area of the detected moving object when the area of the detected moving object is smaller than a predetermined threshold.
  • an environmental image restoration unit a line-of-sight estimating unit that estimates the user's line of sight based on the photographed image; and a line-of-sight estimation unit that acquires sensor data from a sensor included in the image processing device, and calculates the estimated line-of-sight based on the sensor data.
  • a line-of-sight movement detection unit that detects a movement of the user's line of sight starting at
  • a drawing unit that calculates how the AR content looks based on the estimated movement of the line-of-sight
  • a display unit that displays the calculated AR content.
  • an output control section that controls the output so as to perform the control.
  • a mobile terminal or an AR device displays AR content at an accurate position. This makes it possible to accurately present AR content to the user.
  • FIG. 1 is a block diagram showing an example of the hardware configuration of a video processing device according to the first embodiment.
  • FIG. 2 is a block diagram showing the software configuration of the video processing apparatus in the first embodiment in relation to the hardware configuration shown in FIG.
  • FIG. 3 is a flowchart illustrating an example of an operation for the video processing device to display AR content at a correct position in a captured image.
  • FIG. 4 is a diagram showing an example of a photographed image.
  • FIG. 5 is a diagram illustrating an example when a moving object is detected in a photographed image.
  • FIG. 6 is a diagram showing an example of a case where a video of a portion extracted as a moving object is complemented with surrounding images.
  • FIG. 7 is a diagram showing an example of setting AR content corresponding to feature points.
  • FIG. 1 is a block diagram showing an example of the hardware configuration of a video processing device according to the first embodiment.
  • FIG. 2 is a block diagram showing the software configuration of the video processing apparatus in the first embodiment in
  • FIG. 8 is a diagram showing an example of how the calculated AR content looks.
  • FIG. 9 is a diagram illustrating an example of a captured image captured by a camera in the second embodiment.
  • FIG. 10 is a block diagram showing the software configuration of the video processing apparatus in the second embodiment in relation to the hardware configuration shown in FIG. 1.
  • FIG. 11 is a flowchart illustrating an example of an operation for the video processing device to display AR content at a correct position in a captured image.
  • FIG. 12 is a diagram showing an example of detecting unevenness on a road surface in a photographed image.
  • FIG. 13 is a diagram showing an example in which unevenness on a road surface in a photographed image is blurred.
  • FIG. 1 is a block diagram showing an example of the hardware configuration of a video processing device 1 according to the first embodiment.
  • the video processing device 1 is a computer that analyzes input data, generates and outputs output data.
  • the video processing device 1 may be, for example, an AR device including AR glasses, smart glasses, or other wearable devices. That is, the video processing device 1 may be a device worn and used by a user.
  • the video processing device 1 includes a control section 10, a program storage section 20, a data storage section 30, a communication interface 40, and an input/output interface 50.
  • the control unit 10, program storage unit 20, data storage unit 30, communication interface 40, and input/output interface 50 are communicably connected to each other via a bus.
  • the communication interface 40 may be communicably connected to an external device via a network.
  • the input/output interface 50 is communicably connected to the input device 2, the output device 3, the camera 4, and the inertial sensor 5.
  • the control unit 10 controls the video processing device 1.
  • the control unit 10 includes a hardware processor such as a central processing unit (CPU).
  • the control unit 10 may be an integrated circuit capable of executing various programs.
  • the program storage unit 20 includes non-volatile memories that can be written to and read from at any time such as EPROM (Erasable Programmable Read Only Memory), HDD (Hard Disk Drive), and SSD (Solid State Drive), as well as ROM ( It can be used in combination with non-volatile memory such as Read Only Memory).
  • the program storage unit 20 stores programs necessary to execute various processes. That is, the control unit 10 can implement various controls and operations by reading and executing programs stored in the program storage unit 20.
  • the data storage unit 30 is a storage that uses a combination of a non-volatile memory that can be written to and read from at any time, such as an HDD or a memory card, and a volatile memory such as a RAM (Random Access Memory), as a storage medium. .
  • the data storage unit 30 is used to store data acquired and generated while the control unit 10 executes programs and performs various processes.
  • the communication interface 40 includes one or more wired or wireless communication modules.
  • the communication interface 40 includes a communication module that makes a wired or wireless connection to an external device via a network.
  • Communication interface 40 may include a wireless communication module that wirelessly connects to external devices such as Wi-Fi access points and base stations.
  • the communication interface 40 may include a wireless communication module for wirelessly connecting to an external device using short-range wireless technology. That is, the communication interface 40 may be any general communication interface as long as it is capable of communicating with an external device under the control of the control unit 10 and transmitting and receiving various information including past performance data. .
  • the input/output interface 50 is connected to the input device 2, output device 3, camera 4, inertial sensor 5, etc.
  • the input/output interface 50 is an interface that allows information to be transmitted and received between the input device 2, the output device 3, and the plurality of cameras 4 and inertial sensors 5.
  • the input/output interface 50 may be integrated with the communication interface 40.
  • the video processing device 1 and at least one of the input device 2, the output device 3, the camera 4, and the inertial sensor 5 are wirelessly connected using short-range wireless technology or the like. Information may also be sent and received using
  • the input device 2 may include, for example, a keyboard, a pointing device, etc. for the user to input various information including past performance data to the video processing device 1.
  • the input device 2 also includes a reader for reading data to be stored in the program storage section 20 or the data storage section 30 from a memory medium such as a USB memory, and a disk device for reading such data from a disk medium. May be included.
  • the output device 3 includes a display that displays images captured by the camera 4, AR content, and the like.
  • the output device 3 may be integrated with the video processing device 1.
  • the video processing device 1 is AR glasses or smart glasses, the output device 3 is a part of the glasses.
  • the camera 4 is capable of photographing environments such as landscapes, and may be a general camera 4 that can be attached to the video processing device 1.
  • the environment generally refers to the scenery that is photographed.
  • the camera 4 may be integrated with the video processing device 1.
  • the camera 4 may output the captured image to the control unit 10 of the video processing device 1 through the input/output interface 50.
  • the inertial sensor 5 includes, for example, an acceleration sensor, an angular velocity sensor, a geomagnetic sensor, and the like.
  • the inertial sensor 5 senses the moving speed and head movement of the user wearing the AR device, and outputs sensor data according to the sensing to the control unit 10.
  • FIG. 2 is a block diagram showing the software configuration of the video processing device 1 in the first embodiment in relation to the hardware configuration shown in FIG. 1.
  • the control unit 10 includes an image acquisition unit 101 , a moving object detection unit 102 , an environment image restoration unit 103 , a feature point extraction unit 104 , a line of sight estimation unit 105 , a line of sight movement estimation unit 106 , and a sensor data control unit 107 , an AR content drawing section 108 , and an output control section 109 .
  • the image acquisition unit 101 acquires a photographed image taken by the camera 4. Note that the image acquisition unit 101 may store the captured image in the image storage unit 301.
  • the moving object detection unit 102 detects a moving object from the captured image.
  • the moving object detection unit 102 detects the moving object MB appearing in the photographed image.
  • the moving object may be any arbitrary object such as a motorized bicycle, an electric bicycle, a motorcycle, or a vehicle.
  • the moving object may include only the moving object, or may include a part of the body of the user riding the moving object, such as an arm. Further, a general technique may be used as the detection method.
  • the environmental image restoration unit 103 complements the video of the portion detected as a moving object.
  • the environmental image restoration unit 103 replaces the image of the portion detected as a moving object with an environmental image supplemented from the environmental image of the photographed image. Note that a general technique may be used for the interpolation method.
  • the feature point extraction unit 104 extracts things in the environment of the photographed image as feature points. For example, the feature point extracting unit 104 extracts as feature points those located near the feature point space stored in the space storage unit 302, which will be described later.
  • the line of sight estimation unit 105 estimates the position of the line of sight by comparing the feature points extracted by the feature point extraction unit 104 with the feature point space stored in the spatial storage unit 302. Note that details of the method for estimating the position of the line of sight will be described later.
  • the line-of-sight movement estimation unit 106 estimates the line-of-sight movement.
  • the line of sight movement estimating unit 106 moves the three-dimensional movement measured by the sensor data control unit 107 (described later) from the position of the line of sight received from the line of sight estimation unit 105 as a starting point, thereby generating a line of sight that follows the movement of the user's head.
  • Estimate movement That is, the line-of-sight movement estimating unit 106 estimates the movement of the user's line of sight starting from the line-of-sight position estimated by the line-of-sight estimation unit 105 based on the sensor data.
  • the sensor data control unit 107 acquires sensor data from the inertial sensor 5. Then, the sensor data control unit 107 measures the user's head movement, body movement, etc. from the acquired sensor data. For example, the sensor data control unit 107 measures the user's three-dimensional movement (for example, the user's head movement) based on the sensor data.
  • the AR content drawing unit 108 calculates how the AR content looks.
  • the AR content drawing unit 108 draws the estimated line-of-sight movement included in the line-of-sight movement information received from the line-of-sight estimation unit 105, that is, the movement destination line of sight, into the AR content space corresponding to the feature point space stored in the spatial storage unit 302. and calculates how the AR content will appear in the set space.
  • the output control unit 109 outputs AR content information.
  • the output control unit 109 controls the output device 3 to draw AR content.
  • the output control unit 109 controls the adjusted AR content to be displayed on AR glasses or the like.
  • the data storage unit 30 includes an image storage unit 301 and a spatial storage unit 302.
  • the image storage unit 301 may store captured images acquired by the image acquisition unit 101.
  • the captured image stored in the image storage unit 301 may have information about the longitude and latitude in the real world where the captured image was captured, which is acquired by the video processing device 1. Further, the image storage unit 301 may automatically delete the captured image after a predetermined period of time has passed.
  • the space storage unit 302 stores a feature point space and an AR content space corresponding to the feature point space.
  • the feature point space may be located at a preset position in the photographed image, and for example, two positions may be set. Then, an AR content space may be set between the two feature point spaces.
  • the video processing device 1 extracts feature points from the image taken by the camera 4. Furthermore, the video processing device 1 also extracts feature points from a captured image that has been captured in advance (for example, a captured image of the previous frame or several frames before the image) (hereinafter referred to as feature point space). ) and the extracted feature points to estimate the user's line of sight (position and direction).
  • feature point space is a space constructed for a "surrounding space" set based on a predetermined usage scene. Therefore, the captured image is also a captured image of the surrounding space. Therefore, it is assumed that the positional relationships of both the feature point space and the captured image are identified based on the "surrounding space.”
  • the video processing device 1 tracks the movement of the user's head based on the inertial data received from the inertial sensor 5, and tracks the user's line of sight described above in more real time.
  • the video processing device 1 positions the user's line of sight, which is being followed in real time, on an AR content space created in advance, and calculates how the AR content looks from there.
  • the video processing device 1 causes the output device 3 to draw the calculated appearance.
  • the AR system generates AR content and causes the video processing device 1, such as a smartphone or AR glasses, to display the AR content.
  • the video processing device 1 such as a smartphone or AR glasses
  • this method assumes that the camera 4 reflects the environment, so in the case of a captured image that includes both the environment and the mobile object MB, the video processing device 1 can process the AR content. may fail to display in the correct position.
  • the operation of the video processing device 1 for displaying AR content in the correct position even in a photographed image in which the environment and a moving object coexist will be described below.
  • FIG. 3 is a flowchart illustrating an example of an operation by which the video processing device 1 displays AR content at a correct position in a captured image.
  • the operation of this flowchart is realized by the control unit 10 of the video processing device 1 reading and executing the program stored in the program storage unit 20.
  • This operation flow is started, for example, when the user inputs an instruction to display AR content or when a predetermined condition is satisfied, and the control unit 10 outputs an instruction to display AR content.
  • this operation flow may be started when the video processing device 1 is activated and the camera 4 acquires a photographed image. Further, in this operation, it is assumed that the moving object is a bicycle.
  • step ST101 the image acquisition unit 101 acquires a photographed image taken by the camera 4.
  • the image acquisition unit 101 may store the captured image in the image storage unit 301.
  • the photographed image includes the environment and a moving object.
  • the environment may be a general landscape, as described above. Therefore, the environment refers to the part excluding the moving object.
  • FIG. 4 is a diagram showing an example of a photographed image.
  • the image is taken while the user is riding and driving a bicycle, which is a moving object. Therefore, the photographed image includes the moving object and the environment.
  • the camera 4 is a camera 4 included in the AR glasses that are the video processing device 1, and the photographed image is taken by this camera 4.
  • the user's arms, the bicycle handlebars, and wheels are shown with diagonal lines for the sake of simplicity. Furthermore, in the example of FIG. 3, two buildings Bu are shown in the photographed image. Further, it is assumed that the user shown in the example of FIG. 4 is facing forward.
  • the moving object detection unit 102 detects the moving object MB from the captured image.
  • the moving body detection unit 102 may perform object detection using a general method to detect the moving body MB.
  • the object detection method an object detection method as disclosed in Non-Patent Document 2 may be used. Therefore, a detailed explanation of the object detection method will be omitted here.
  • FIG. 5 is a diagram illustrating an example when a moving body MB is detected in a photographed image.
  • a bicycle and a user are detected as mobile objects MB. That is, the portion detected as the mobile body MB includes the user's arm in addition to the bicycle handle and wheels.
  • the environment image restoration unit 103 complements the video of the portion detected as the mobile object MB.
  • the environment image restoration unit 103 complements the image of the surrounding environment in the video of the portion detected as the mobile object MB. That is, the environmental image restoration unit 103 reproduces the peripheral image of the portion detected as the mobile object MB.
  • the environment image restoration unit 103 may perform complementation using a general method.
  • the environmental image restoration unit 103 may use a complementary technique as disclosed in Non-Patent Document 3. Therefore, a detailed explanation of the complementary technology will be omitted here.
  • FIG. 6 is a diagram showing an example of a case where a video of a portion extracted as a mobile object MB is complemented with surrounding images. As shown in FIG. 6, by complementing the image of the mobile body MB with the surrounding image, the environmental image restoration unit 103 can obtain a photographed image that would have been taken if the mobile body MB did not exist.
  • the line of sight estimation unit 105 estimates the user's line of sight.
  • the feature point extraction unit 104 extracts feature points from the captured image.
  • the feature points may be, for example, two buildings Bu.
  • FIG. 7 is a diagram showing an example of setting AR content corresponding to feature points.
  • the space storage unit 302 presets a "feature point space” with two buildings Bu as feature points, and places a smiley face that is AR content in the middle part of the two buildings Bu in the space. It is assumed that an "AR content space” in which marks are placed is also stored. That is, the space storage unit 302 stores a feature point space and an AR content space corresponding to the feature point space.
  • the AR content is indicated by reference symbol ARC.
  • FIG. 7 is just an example, and the space storage unit 302 may store the AR content space together with various feature point spaces.
  • the feature point extraction unit 104 looks over the entire photographed image and extracts specific feature points such as boundaries (edges) of objects or corners of objects.
  • the thing may be a boundary between objects.
  • FIG. 7 it is possible to extract the boundary of the building Bu, which is a specific feature point.
  • the line of sight estimation unit 105 estimates the position of the line of sight by comparing the feature points extracted by the feature point extraction unit 104 with the feature point space stored in the spatial storage unit 302.
  • the line of sight estimation unit 105 may estimate the line of sight using vision-based AR technology or the like.
  • the vision-based AR technology may be a general technology such as PTAM, SmartAR, Microsoft Hololens, etc., which are markerless AR technologies, for example. Therefore, a detailed explanation of the AR technology will be omitted here.
  • the line-of-sight estimation unit 105 outputs the estimated line-of-sight position to the line-of-sight movement estimation unit 106.
  • the line-of-sight estimation unit 105 can perform high-speed processing by comparing feature points extracted and simplified as feature points with the feature point space instead of the image itself.
  • the environmental image restoration unit 103 can complement the portion detected as the moving object (image restoration accuracy is high).
  • the line of sight estimating unit 105 successfully estimates the line of sight.
  • the line-of-sight movement estimation unit 106 estimates the line-of-sight movement.
  • the sensor data control unit 107 acquires sensor data from the inertial sensor 5. Then, the sensor data control unit 107 measures the user's head movement, body movement, etc. from the acquired sensor data.
  • the inertial sensor 5 is an inertial measurement unit (IMU), and the sensor data control unit 107 acquires sensor data such as acceleration, angular velocity, and geomagnetism from the inertial sensor 5.
  • the sensor data control unit 107 may then measure the user's three-dimensional movement (for example, the user's head movement) based on these data. Then, the sensor data control unit 107 outputs the measurement result to the line of sight movement estimation unit 106.
  • IMU inertial measurement unit
  • the line-of-sight movement estimating unit 106 moves the three-dimensional movement measured by the sensor data control unit 107 from the position of the line-of-sight received from the line-of-sight estimation unit 105 as a starting point, thereby generating a line-of-sight movement that follows the movement of the user's head.
  • the line-of-sight movement estimating unit 106 estimates the movement of the user's line of sight starting from the line-of-sight position estimated by the line-of-sight estimation unit 105 based on the sensor data.
  • the eye movement estimation unit 106 outputs eye movement information including the estimated eye movement to the AR content drawing unit 108.
  • the AR content drawing unit 108 calculates how the AR content looks.
  • the AR content drawing unit 108 draws the estimated line-of-sight movement included in the line-of-sight movement information received from the line-of-sight estimation unit 105, that is, the movement destination line of sight, into the AR content space corresponding to the feature point space stored in the spatial storage unit 302. and calculates how the AR content will appear in the set space.
  • FIG. 8 is a diagram showing an example of how the calculated AR content looks.
  • the AR content drawing unit 108 adjusts the appearance of the AR content based on the movement information of the eye movement, and sends AR content information to the output control unit 109 for drawing the adjusted AR content ARC. Output.
  • step ST107 the output control unit 109 outputs AR content information.
  • the output control unit 109 controls the output device 3 to draw the AR content ARC.
  • the output control unit 109 controls the adjusted AR content ARC to be displayed on AR glasses or the like.
  • the video processing device 1 can accurately process AR content even when a portion showing the environment and a portion showing the mobile object MB coexist in an image taken by the camera 4.
  • the ARC can be presented to the user.
  • FIG. 9 is a diagram showing an example of a captured image captured by the camera 4 in the second embodiment. As shown in FIG. 9, when the mobile object MB occupies most of the photographed image, the complementation processing may fail even if the processing in the first embodiment is performed.
  • the hardware configuration of the video processing device 1 in the second embodiment may be the same as the hardware configuration in the first embodiment, so a redundant explanation here will be omitted.
  • FIG. 10 is a block diagram showing the software configuration of the video processing device 1 in the second embodiment in relation to the hardware configuration shown in FIG. 1.
  • the control unit 10 differs from the control unit 10 in the first embodiment in that it includes a moving object area calculation unit 110 and a road image analysis unit 111.
  • the moving body area calculation unit 110 calculates the area of the moving body MB.
  • the moving object area calculation unit 110 may calculate the area of the moving object MB, or may calculate the ratio of the moving object MB to the captured image. Furthermore, the moving object area calculation unit 110 determines whether the calculated area of the moving object MB is equal to or larger than a threshold value.
  • the road surface image analysis unit 111 detects a portion where the unevenness in the environment shown in the photographed image has been deformed into a vertically elongated shape due to movement within the interval of the shutter speed of the camera 4.
  • the road surface image analysis unit 111 may detect irregularities (for example, pebbles, etc.) appearing in the environment of the photographed image using a general method, and detect deformation of the pebbles that are the irregularities. Furthermore, the road surface image analysis unit 111 estimates the moving speed of the mobile body MB from the degree of deformation. Details of the method for estimating the moving speed will be described later.
  • the data storage unit 30 differs from the first embodiment in that it includes a line-of-sight storage unit 303.
  • the line of sight storage unit 303 may store information about the line of sight estimated by the line of sight estimation unit 105. Note that the stored information regarding the line of sight may be deleted after a certain period of time has passed.
  • FIG. 11 is a flowchart illustrating an example of an operation by which the video processing device 1 displays AR content at a correct position in a captured image. The operation of this flowchart is realized by the control unit 10 of the video processing device 1 reading and executing the program stored in the program storage unit 20.
  • This operation flow is started, for example, when the user inputs an instruction to display AR content or when a predetermined condition is satisfied, and the control unit 10 outputs an instruction to display AR content.
  • this operation flow may be started when the video processing device 1 is activated and the camera 4 acquires a photographed image.
  • Step ST201 and step ST202 may be the same as step ST101 and step ST102 described with reference to FIG. 3, so duplicate explanation here will be omitted.
  • the moving object detection section 102 may output the captured image and information about the detected moving object MB to the moving object area calculation section 110.
  • step ST203 the moving body area calculation unit 110 calculates the area of the moving body MB.
  • the moving object area calculation unit 110 may calculate the area of the moving object MB, or may calculate the proportion of the moving object MB in the captured image.
  • step ST204 the moving object area calculation unit 110 determines whether the calculated area of the moving object MB is greater than or equal to a threshold value. If it is determined that the calculated area of the moving body MB is not equal to or larger than the predetermined threshold, that is, if the area of the moving body MB does not occupy most of the captured image, the process proceeds to step ST205. On the other hand, if it is determined that the calculated area is equal to or greater than the predetermined threshold, that is, if the area of the moving body MB occupies most of the captured image, the process proceeds to step ST207. Note that when calculating the ratio, the moving object area calculation unit 110 may determine whether the ratio of the moving object MB is equal to or greater than a threshold value.
  • Step ST205 and step ST206 may be the same as step ST103 and step ST104 described with reference to FIG. 3, so a duplicate description here will be omitted.
  • the line-of-sight estimating unit 105 may store information about the estimated line-of-sight estimation in the line-of-sight storage unit 303 together with time information.
  • the process estimates the user's line of sight using a method that allows estimation of the line of sight even when the user looks down.
  • the road image analysis unit 111 detects a portion where the unevenness in the environment shown in the photographed image has been deformed into a vertically elongated shape due to movement within the interval of the shutter speed of the camera 4.
  • the road surface image analysis unit 111 may detect irregularities (for example, pebbles, etc.) appearing in the environment of the photographed image using a general method, and detect deformation of the pebbles that are the irregularities. For example, irregularities (pebbles) are detected as a rectangle circumscribing the irregularities. Therefore, the road image analysis unit 111 estimates the degree of deformation from the ratio of the length and width of this rectangle. Then, the road image analysis unit 111 estimates the moving speed of the mobile body MB from the degree of deformation.
  • irregularities for example, pebbles, etc.
  • the road surface video analysis unit 111 estimates the speed of the moving object MB shown in the photographed image based on, for example, the intensity of blur, the mounting angle of the camera 4, and the like.
  • a method for estimating the speed of the mobile body MB a general technique may be used.
  • a method for estimating the speed of the mobile body MB a method for estimating the speed of the mobile body MB as disclosed in Non-Patent Document 4 may be used. Therefore, a detailed explanation of the method for estimating the speed of the mobile body MB will be omitted here.
  • FIG. 12 is a diagram showing an example of detecting unevenness on a road surface in a photographed image.
  • unevenness on the road surface is indicated by reference symbol CC.
  • the road image analysis unit 111 detects irregularities such as pebbles on the road surface from the photographed image.
  • FIG. 13 is a diagram showing an example in which unevenness on a road surface in a photographed image is blurred. As shown in FIG. 13, as the speed of the moving body MB increases, the detected pebbles become blurred.
  • the road surface image analysis unit 111 may estimate the speed of the moving body MB using the degree of blur.
  • the line of sight estimation unit 105 estimates the line of sight.
  • the line-of-sight estimation unit 105 acquires information about the previously stored line-of-sight estimation from the current time from the line-of-sight storage unit 303, and estimates the current line-of-sight position based on the result of the line-of-sight estimation and the movement speed.
  • Steps ST209 to ST211 may be the same as steps ST105 to ST107 described with reference to FIG. 3, so a duplicate explanation here will be omitted.
  • the AR content space is linked to a coordinate system such as longitude and latitude in the real world. Therefore, even if the line of sight estimation method is a method that utilizes the distortion of unevenness in the captured image, there are two points in the AR content space: the moving distance calculated from speed and time, and the longitude/latitude coordinate system. It is possible to link and map each other based on the distance between them. Therefore, through the processing in steps ST209 to ST211, the video processing device 1 can correctly recognize the AR content space and can accurately present the AR content ARC to the user.
  • the video processing device 1 can accurately perform AR processing even when the portion of the image captured by the camera 4 is dominated by the portion of the mobile object MB rather than the portion of the environment.
  • the content ARC can be presented to the user.
  • the photographed image is not limited to the camera 4 included in the video processing device 1.
  • it may be an independent camera 4 connected to the video processing device 1.
  • the camera 4 is installed at a location (for example, above the user's head) where a captured image from which the user's line of sight can be estimated can be captured.
  • the method described in the above embodiments can be applied to, for example, magnetic disks (floppy (registered trademark) disks, hard disks, etc.), optical disks (CD-ROMs, DVDs, etc.) as programs (software means) that can be executed by a computer. , MO, etc.), semiconductor memory (ROM, RAM, flash memory, etc.), and can also be transmitted and distributed via a communication medium.
  • the programs stored on the medium side also include a setting program for configuring software means (including not only execution programs but also tables and data structures) in the computer to be executed by the computer.
  • a computer that realizes this device reads a program stored in a storage medium, and if necessary, constructs software means using a setting program, and executes the above-described processing by controlling the operation of the software means.
  • the storage medium referred to in this specification is not limited to those for distribution, and includes storage media such as magnetic disks and semiconductor memories provided inside computers or devices connected via a network.
  • the present invention is not limited to the above-described embodiments, and various modifications can be made at the implementation stage without departing from the spirit thereof. Moreover, each embodiment may be implemented by appropriately combining them as much as possible, and in that case, the combined effects can be obtained. Further, the embodiments described above include inventions at various stages, and various inventions can be extracted by appropriately combining the plurality of disclosed constituent elements.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Graphics (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

A video processing device according to one embodiment is to be worn by a user and comprises an image acquisition unit that acquires a captured image that has been captured by a camera and includes a mobile body being ridden by the user and the surroundings, a mobile body detection unit that detects the mobile body from the captured image, a surroundings image reconstruction unit that supplements an image of the surroundings at the portion of the detected mobile body when the area of the detected mobile body is smaller than a prescribed threshold value, a line of sight estimation unit that estimates the line of sight of the user on the basis of the captured image, a line of sight movement detection unit that acquires sensor data from a sensor of the video processing device and detects movement of the line of sight of the user from the estimated line of sight on the basis of the sensor data, a rendering unit that calculates the appearance of AR content on the basis of the movement of the estimated line of sight, and an output control unit that performs control to display the calculated AR content.

Description

映像処理装置、映像処理方法、および映像処理プログラムVideo processing device, video processing method, and video processing program
 この発明は、映像処理装置、映像処理方法、および映像処理プログラムに関する。 The present invention relates to a video processing device, a video processing method, and a video processing program.
 拡張現実(AR:Augmented Reality)システムを利用するユーザは、携帯端末またはARデバイス越しに現実世界の実空間を見ることが可能である。この際、実空間に付加情報として、ナビゲーション情報または3Dデータ等のコンテンツ(以下、ARコンテンツと記載する)が提示される。すなわち、ARシステムを利用するユーザは、現実世界にARコンテンツが重ねて表示されており、このコンテンツの情報を利用することができる。 A user using an augmented reality (AR) system can view the real space of the real world through a mobile terminal or an AR device. At this time, content such as navigation information or 3D data (hereinafter referred to as AR content) is presented as additional information in the real space. That is, a user using an AR system can see AR content superimposed on the real world and use information about this content.
 例えば、ARシステムを利用するユーザが移動体に乗って移動している場合、ARデバイスにより映し出される映像は、カメラの映像に環境を映した部分(自転車の前方の風景)と、移動体を映した部分(自転車の車体の一部)とが混在する。 For example, when a user using an AR system is moving around on a moving object, the image displayed by the AR device will include a portion of the camera image showing the environment (scenery in front of the bicycle) and a portion showing the moving object. (a part of the bicycle body).
 従来の自己位置推定処理は、カメラの映像が環境を映したものであることを前提としており、正確に処理できない。その結果、ARデバイスは、正しい位置にARコンテンツを表示することができないという問題がある。 Conventional self-position estimation processing is based on the premise that the camera image reflects the environment, and cannot be processed accurately. As a result, there is a problem in that the AR device cannot display AR content in the correct position.
 この発明は上記事情に着目してなされたもので、その目的とするところは、カメラの映像に環境を映した部分と移動体を映した部分とが混在した場合であっても、携帯端末またはARデバイスは、正確な位置にARコンテンツを表示することができる技術を提供することにある。 This invention has been made in view of the above-mentioned circumstances, and its purpose is to prevent mobile terminals or The purpose of an AR device is to provide a technology that can display AR content at a precise location.
 上記課題を解決するためにこの発明の一態様は、ユーザが装着する映像処理装置であって、カメラで撮影した、前記ユーザが乗っている移動体と環境とを含む撮影画像を取得する画像取得部と、前記撮影画像から前記移動体を検出する移動体検出部と、前記検出された移動体の面積が所定の閾値より小さい場合、前記検出された移動体の部分における環境の画像を補完する環境画像復元部と、前記撮影画像に基づいて前記ユーザの視線を推定する視線推定部と、前記映像処理装置が備えるセンサからセンサデータを取得し、前記センサデータに基づいて、前記推定された視線を起点とした前記ユーザの視線の移動を検出する視線移動検出部と、前記推定された視線の移動に基づいて、ARコンテンツの見え方を算出する描画部と、前記算出されたARコンテンツを表示するように制御する出力制御部と、を備えるようにしたものである。 In order to solve the above problems, one aspect of the present invention is an image processing device worn by a user, which acquires an image captured by a camera and including a moving object on which the user is riding and an environment. a moving object detecting section that detects the moving object from the photographed image; and a moving object detecting section that detects the moving object from the photographed image, and complements an image of the environment in the area of the detected moving object when the area of the detected moving object is smaller than a predetermined threshold. an environmental image restoration unit; a line-of-sight estimating unit that estimates the user's line of sight based on the photographed image; and a line-of-sight estimation unit that acquires sensor data from a sensor included in the image processing device, and calculates the estimated line-of-sight based on the sensor data. a line-of-sight movement detection unit that detects a movement of the user's line of sight starting at , a drawing unit that calculates how the AR content looks based on the estimated movement of the line-of-sight, and a display unit that displays the calculated AR content. and an output control section that controls the output so as to perform the control.
 この発明の一態様によれば、カメラの映像に環境を映した部分と移動体を映した部分とが混在した場合であっても、携帯端末またはARデバイスは、正確な位置にARコンテンツを表示することができ、これにより、ユーザにARコンテンツを正確に提示することが可能となる。 According to one aspect of the present invention, even when a camera image includes a portion showing the environment and a portion showing a moving object, a mobile terminal or an AR device displays AR content at an accurate position. This makes it possible to accurately present AR content to the user.
図1は、第1の実施形態に係る映像処理装置のハードウェア構成の一例を示すブロック図である。FIG. 1 is a block diagram showing an example of the hardware configuration of a video processing device according to the first embodiment. 図2は、第1の実施形態における映像処理装置のソフトウェア構成を、図1に示したハードウェア構成に関連付けて示すブロック図である。FIG. 2 is a block diagram showing the software configuration of the video processing apparatus in the first embodiment in relation to the hardware configuration shown in FIG. 図3は、映像処理装置が撮影画像の正しい位置にARコンテンツを表示させるための動作の一例を示すフローチャートである。FIG. 3 is a flowchart illustrating an example of an operation for the video processing device to display AR content at a correct position in a captured image. 図4は、撮影画像の一例を示す図である。FIG. 4 is a diagram showing an example of a photographed image. 図5は、撮影画像において移動体が検出した際の一例を示す図である。FIG. 5 is a diagram illustrating an example when a moving object is detected in a photographed image. 図6は、移動体として抽出された部分の映像を周辺画像で補完した場合の一例を示した図である。FIG. 6 is a diagram showing an example of a case where a video of a portion extracted as a moving object is complemented with surrounding images. 図7は、特徴点に対応するARコンテンツの設定例を示した図である。FIG. 7 is a diagram showing an example of setting AR content corresponding to feature points. 図8は、算出されたARコンテンツの見え方の一例を示した図である。FIG. 8 is a diagram showing an example of how the calculated AR content looks. 図9は、第2の実施形態におけるカメラで撮影された撮影画像の一例を示す図である。FIG. 9 is a diagram illustrating an example of a captured image captured by a camera in the second embodiment. 図10は、第2の実施形態における映像処理装置のソフトウェア構成を、図1に示したハードウェア構成に関連付けて示すブロック図である。FIG. 10 is a block diagram showing the software configuration of the video processing apparatus in the second embodiment in relation to the hardware configuration shown in FIG. 1. 図11は、映像処理装置が撮影画像の正しい位置にARコンテンツを表示させるための動作の一例を示すフローチャートである。FIG. 11 is a flowchart illustrating an example of an operation for the video processing device to display AR content at a correct position in a captured image. 図12は、撮影画像における路面上の凹凸を検出した一例を示した図である。FIG. 12 is a diagram showing an example of detecting unevenness on a road surface in a photographed image. 図13は、撮影画像における路面上の凹凸に対してブラーが掛かっている一例を示した図である。FIG. 13 is a diagram showing an example in which unevenness on a road surface in a photographed image is blurred.
 以下、図面を参照してこの発明に係る実施形態を説明する。なお、以降、説明済みの要素と同一または類似の要素には同一または類似の符号を付し、重複する説明については基本的に省略する。例えば、複数の同一または類似の要素が存在する場合に、各要素を区別せずに説明するために共通の符号を用いることがあるし、各要素を区別して説明するために当該共通の符号に加えて枝番号を用いることもある。 Hereinafter, embodiments according to the present invention will be described with reference to the drawings. Note that, hereinafter, elements that are the same or similar to elements that have already been explained will be given the same or similar numerals, and overlapping explanations will basically be omitted. For example, when there are multiple identical or similar elements, a common code may be used to explain each element without distinction, or a common code may be used to distinguish and explain each element. In addition, branch numbers may also be used.
 [第1の実施形態] 
 (構成) 
 図1は、第1の実施形態に係る映像処理装置1のハードウェア構成の一例を示すブロック図である。 
 映像処理装置1は、入力されたデータを解析して、出力データを生成し出力する、コンピュータである。映像処理装置1は、例えば、ARグラス、スマートグラス、または他のウェアラブルデバイスを含むARデバイスであって良い。すなわち、映像処理装置1は、ユーザが装着して使用するデバイスであって良い。
[First embodiment]
(composition)
FIG. 1 is a block diagram showing an example of the hardware configuration of a video processing device 1 according to the first embodiment.
The video processing device 1 is a computer that analyzes input data, generates and outputs output data. The video processing device 1 may be, for example, an AR device including AR glasses, smart glasses, or other wearable devices. That is, the video processing device 1 may be a device worn and used by a user.
 図1に示すように、映像処理装置1は、制御部10、プログラム記憶部20、データ記憶部30、通信インタフェース40、および入出力インタフェース50を備える。制御部10、プログラム記憶部20、データ記憶部30、通信インタフェース40、および入出力インタフェース50は、バスを介して互いに通信可能に接続されている。さらに通信インタフェース40は、ネットワークを介して外部装置と通信可能に接続されてよい。また、入出力インタフェース50は、入力装置2、出力装置3、カメラ4、および慣性センサ5と通信可能に接続される。 As shown in FIG. 1, the video processing device 1 includes a control section 10, a program storage section 20, a data storage section 30, a communication interface 40, and an input/output interface 50. The control unit 10, program storage unit 20, data storage unit 30, communication interface 40, and input/output interface 50 are communicably connected to each other via a bus. Further, the communication interface 40 may be communicably connected to an external device via a network. Further, the input/output interface 50 is communicably connected to the input device 2, the output device 3, the camera 4, and the inertial sensor 5.
 制御部10は、映像処理装置1を制御する。制御部10は、中央処理ユニット(CPU:Central Processing Unit)等のハードウェアプロセッサを備える。例えば、制御部10は、様々なプログラムを実行することが可能な集積回路であっても良い。 The control unit 10 controls the video processing device 1. The control unit 10 includes a hardware processor such as a central processing unit (CPU). For example, the control unit 10 may be an integrated circuit capable of executing various programs.
 プログラム記憶部20は、記憶媒体として、例えば、EPROM(Erasable Programmable Read Only Memory)、HDD(Hard Disk Drive)、SSD(Solid State Drive)等の随時書込みおよび読出しが可能な不揮発性メモリと、ROM(Read Only Memory)等の不揮発性メモリとを組み合わせて使用することができる。プログラム記憶部20は、各種処理を実行するために必要なプログラムを格納している。すなわち、制御部10は、プログラム記憶部20に格納されたプログラムを読み出して実行することにより各種制御および動作を実現し得る。 The program storage unit 20 includes non-volatile memories that can be written to and read from at any time such as EPROM (Erasable Programmable Read Only Memory), HDD (Hard Disk Drive), and SSD (Solid State Drive), as well as ROM ( It can be used in combination with non-volatile memory such as Read Only Memory). The program storage unit 20 stores programs necessary to execute various processes. That is, the control unit 10 can implement various controls and operations by reading and executing programs stored in the program storage unit 20.
 データ記憶部30は、記憶媒体として、例えば、HDD、メモリカード等の随時書込みおよび読出しが可能な不揮発性メモリと、RAM(Random Access Memory)等の揮発性メモリとを組み合わせて使用したストレージである。データ記憶部30は、制御部10がプログラムを実行して各種処理を行う過程で取得および生成されたデータを記憶するために用いられる。 The data storage unit 30 is a storage that uses a combination of a non-volatile memory that can be written to and read from at any time, such as an HDD or a memory card, and a volatile memory such as a RAM (Random Access Memory), as a storage medium. . The data storage unit 30 is used to store data acquired and generated while the control unit 10 executes programs and performs various processes.
 通信インタフェース40は、1つ以上の有線または無線の通信モジュールを含む。例えば、通信インタフェース40は、ネットワークを介して外部装置と有線または無線接続する通信モジュールを含む。通信インタフェース40は、Wi-Fiアクセスポイントおよび基地局等の外部装置と無線接続する無線通信モジュールを含んでも良い。さらに、通信インタフェース40は、近距離無線技術を利用して外部装置と無線接続するための無線通信モジュールを含んでも良い。すなわち、通信インタフェース40は、制御部10の制御の下、外部装置との間で通信を行い、過去の実績データを含む各種情報を送受信することができるものであれば一般的な通信インタフェースで良い。 The communication interface 40 includes one or more wired or wireless communication modules. For example, the communication interface 40 includes a communication module that makes a wired or wireless connection to an external device via a network. Communication interface 40 may include a wireless communication module that wirelessly connects to external devices such as Wi-Fi access points and base stations. Furthermore, the communication interface 40 may include a wireless communication module for wirelessly connecting to an external device using short-range wireless technology. That is, the communication interface 40 may be any general communication interface as long as it is capable of communicating with an external device under the control of the control unit 10 and transmitting and receiving various information including past performance data. .
 入出力インタフェース50は、入力装置2、出力装置3、カメラ4、慣性センサ5等と接続される。入出力インタフェース50は、入力装置2、出力装置3、および複数のカメラ4、慣性センサ5との間で情報の送受信を可能にするインタフェースである。入出力インタフェース50は、通信インタフェース40と一体であってもよい。例えば、映像処理装置1と、入力装置2、出力装置3、カメラ4、慣性センサ5の少なくとも1つとは、近距離無線技術等を使用して無線接続されており、当該近距離無線技術を用いて情報の送受信を行ってもよい。 The input/output interface 50 is connected to the input device 2, output device 3, camera 4, inertial sensor 5, etc. The input/output interface 50 is an interface that allows information to be transmitted and received between the input device 2, the output device 3, and the plurality of cameras 4 and inertial sensors 5. The input/output interface 50 may be integrated with the communication interface 40. For example, the video processing device 1 and at least one of the input device 2, the output device 3, the camera 4, and the inertial sensor 5 are wirelessly connected using short-range wireless technology or the like. Information may also be sent and received using
 入力装置2は、例えば、ユーザが映像処理装置1に対して過去の実績データを含む各種情報を入力するためのキーボードやポインティングデバイス等を含んでも良い。また、入力装置2は、プログラム記憶部20またはデータ記憶部30に格納するべきデータを、USBメモリ等のメモリ媒体から読み出すためのリーダや、そのようなデータをディスク媒体から読み出すためのディスク装置を含んでも良い。 The input device 2 may include, for example, a keyboard, a pointing device, etc. for the user to input various information including past performance data to the video processing device 1. The input device 2 also includes a reader for reading data to be stored in the program storage section 20 or the data storage section 30 from a memory medium such as a USB memory, and a disk device for reading such data from a disk medium. May be included.
 出力装置3は、カメラ4で撮影した映像、ARコンテンツを表示するディスプレイ等を含む。出力装置3は、映像処理装置1と一体になっていても良い。例えば、映像処理装置1がARグラスまたはスマートグラス等である場合、出力装置3は、グラスの部分になる。 The output device 3 includes a display that displays images captured by the camera 4, AR content, and the like. The output device 3 may be integrated with the video processing device 1. For example, when the video processing device 1 is AR glasses or smart glasses, the output device 3 is a part of the glasses.
 カメラ4は、風景等の環境を撮影することが可能であり、映像処理装置1に装着可能な一般的なカメラ4であって良い。ここで、環境は、一般的に撮影される風景を指す。カメラ4は、映像処理装置1と一体になっていても良い。カメラ4は、撮影した撮影画像を入出力インタフェース50を通じて、映像処理装置1の制御部10に出力して良い。 The camera 4 is capable of photographing environments such as landscapes, and may be a general camera 4 that can be attached to the video processing device 1. Here, the environment generally refers to the scenery that is photographed. The camera 4 may be integrated with the video processing device 1. The camera 4 may output the captured image to the control unit 10 of the video processing device 1 through the input/output interface 50.
 慣性センサ5は、例えば、加速度センサ、角速度センサ、地磁気センサ等を含む。例えば、映像処理装置1がARデバイスである場合、慣性センサ5は、ARデバイスを装着したユーザの移動スピード、頭の動きを感知し、感知に応じたセンサデータを制御部10に出力する。 The inertial sensor 5 includes, for example, an acceleration sensor, an angular velocity sensor, a geomagnetic sensor, and the like. For example, when the video processing device 1 is an AR device, the inertial sensor 5 senses the moving speed and head movement of the user wearing the AR device, and outputs sensor data according to the sensing to the control unit 10.
 図2は、第1の実施形態における映像処理装置1のソフトウェア構成を、図1に示したハードウェア構成に関連付けて示すブロック図である。
 制御部10は、画像取得部101と、移動体検出部102と、環境画像復元部103と、特徴点抽出部104と、視線推定部105と、視線移動推定部106と、センサデータ制御部107と、ARコンテンツ描画部108と、出力制御部109と、を備える。
FIG. 2 is a block diagram showing the software configuration of the video processing device 1 in the first embodiment in relation to the hardware configuration shown in FIG. 1.
The control unit 10 includes an image acquisition unit 101 , a moving object detection unit 102 , an environment image restoration unit 103 , a feature point extraction unit 104 , a line of sight estimation unit 105 , a line of sight movement estimation unit 106 , and a sensor data control unit 107 , an AR content drawing section 108 , and an output control section 109 .
 画像取得部101は、カメラ4が撮影した撮影画像を取得する。なお、画像取得部101は、撮影画像を画像記憶部301に記憶させて良い。 The image acquisition unit 101 acquires a photographed image taken by the camera 4. Note that the image acquisition unit 101 may store the captured image in the image storage unit 301.
 移動体検出部102は、撮影画像から移動体を検出する。移動体検出部102は、撮影画像に写った移動体MBを検出する。移動体は、原動機付き自転車、電動機付き自転車、自動二輪車、車両等任意のものであって良いのは勿論である。さらに、移動体は、移動体のみを含むだけでも良いし、移動体に乗っているユーザの体の一部、例えば、腕等も含んで良い。また、検出方法は、一般的な技術を用いて良い。 The moving object detection unit 102 detects a moving object from the captured image. The moving object detection unit 102 detects the moving object MB appearing in the photographed image. Of course, the moving object may be any arbitrary object such as a motorized bicycle, an electric bicycle, a motorcycle, or a vehicle. Further, the moving object may include only the moving object, or may include a part of the body of the user riding the moving object, such as an arm. Further, a general technique may be used as the detection method.
 環境画像復元部103は、移動体として検出された部分の映像を補完する。環境画像復元部103は、移動体として検出された部分の画像を撮影画像の環境の画像から補完された環境の画像に置き換える。なお、補完の方法は、一般的な技術を用いて良い。 The environmental image restoration unit 103 complements the video of the portion detected as a moving object. The environmental image restoration unit 103 replaces the image of the portion detected as a moving object with an environmental image supplemented from the environmental image of the photographed image. Note that a general technique may be used for the interpolation method.
 特徴点抽出部104は、撮影画像の環境内にあるものを特徴点として抽出する。例えば、特徴点抽出部104は、後述する空間記憶部302に記憶された特徴点空間付近にあるものを特徴点として抽出する。 The feature point extraction unit 104 extracts things in the environment of the photographed image as feature points. For example, the feature point extracting unit 104 extracts as feature points those located near the feature point space stored in the space storage unit 302, which will be described later.
 視線推定部105は、特徴点抽出部104が抽出した特徴点を空間記憶部302に記憶された特徴点空間と突合することにより、視線の位置を推定する。なお、視線の位置の推定方法の詳細は後述する。 The line of sight estimation unit 105 estimates the position of the line of sight by comparing the feature points extracted by the feature point extraction unit 104 with the feature point space stored in the spatial storage unit 302. Note that details of the method for estimating the position of the line of sight will be described later.
 視線移動推定部106は、視線移動を推定する。視線移動推定部106は、後述するセンサデータ制御部107により計測された3次元の動きを視線推定部105から受信した視線の位置を起点に移動させることにより、ユーザの頭の動きに追従した視線移動を推定する。すなわち、視線移動推定部106は、センサデータに基づいて、視線推定部105により推定された視線の位置を起点としたユーザの視線の移動を推定する。 The line-of-sight movement estimation unit 106 estimates the line-of-sight movement. The line of sight movement estimating unit 106 moves the three-dimensional movement measured by the sensor data control unit 107 (described later) from the position of the line of sight received from the line of sight estimation unit 105 as a starting point, thereby generating a line of sight that follows the movement of the user's head. Estimate movement. That is, the line-of-sight movement estimating unit 106 estimates the movement of the user's line of sight starting from the line-of-sight position estimated by the line-of-sight estimation unit 105 based on the sensor data.
 センサデータ制御部107は、慣性センサ5からセンサデータを取得する。そしてセンサデータ制御部107は、取得したセンサデータからユーザの頭の動き、体の動き等を計測する。例えば、センサデータ制御部107は、センサデータに基づいてユーザの3次元の動き(例えば、ユーザの頭の動き)を計測する。 The sensor data control unit 107 acquires sensor data from the inertial sensor 5. Then, the sensor data control unit 107 measures the user's head movement, body movement, etc. from the acquired sensor data. For example, the sensor data control unit 107 measures the user's three-dimensional movement (for example, the user's head movement) based on the sensor data.
 ARコンテンツ描画部108は、ARコンテンツの見え方を算出する。ARコンテンツ描画部108は、視線推定部105から受信した視線移動情報に含まれる推定した視線移動、すなわち、移動先の視線を空間記憶部302に記憶された特徴点空間に対応するARコンテンツ空間に設定し、設定された空間上でのARコンテンツの見え方を算出する。 The AR content drawing unit 108 calculates how the AR content looks. The AR content drawing unit 108 draws the estimated line-of-sight movement included in the line-of-sight movement information received from the line-of-sight estimation unit 105, that is, the movement destination line of sight, into the AR content space corresponding to the feature point space stored in the spatial storage unit 302. and calculates how the AR content will appear in the set space.
 出力制御部109は、ARコンテンツ情報を出力する。出力制御部109は、出力装置3にARコンテンツを描画するように制御する。例えば、出力制御部109は、調整されたARコンテンツをARグラス等に表示するように制御する。 The output control unit 109 outputs AR content information. The output control unit 109 controls the output device 3 to draw AR content. For example, the output control unit 109 controls the adjusted AR content to be displayed on AR glasses or the like.
 データ記憶部30は、画像記憶部301と、空間記憶部302と、を備える。 The data storage unit 30 includes an image storage unit 301 and a spatial storage unit 302.
 画像記憶部301は、画像取得部101が取得した撮影画像を記憶して良い。ここで、画像記憶部301に記憶される撮影画像は、映像処理装置1により取得された、撮影画像が撮影された現実世界における経度および緯度についての情報を有していても良い。また、画像記憶部301は、撮影した画像を所定の時間経過後、自動的に削除して良い。 The image storage unit 301 may store captured images acquired by the image acquisition unit 101. Here, the captured image stored in the image storage unit 301 may have information about the longitude and latitude in the real world where the captured image was captured, which is acquired by the video processing device 1. Further, the image storage unit 301 may automatically delete the captured image after a predetermined period of time has passed.
 空間記憶部302は、特徴点空間と当該特徴点空間に対応したARコンテンツ空間とを記憶している。特徴点空間は、撮影画像中の予め設定された位置にあって良く、例えば、2つの位置が設定されていても良い。そして、2つの特徴点空間の中間部にARコンテンツ空間が設定されて良い。 The space storage unit 302 stores a feature point space and an AR content space corresponding to the feature point space. The feature point space may be located at a preset position in the photographed image, and for example, two positions may be set. Then, an AR content space may be set between the two feature point spaces.
 (動作) 
 最初に、一般的なARシステムにおいて、ユーザが使用する映像処理装置1(携帯端末またはARデバイス)にARコンテンツを表示する方法について説明する。
(motion)
First, a method for displaying AR content on the video processing device 1 (mobile terminal or AR device) used by the user in a general AR system will be described.
 映像処理装置1は、カメラ4が撮影した撮影画像から、特徴点を抽出する。さらに、映像処理装置1は、予め撮影された撮影画像(例えば、1つ前のフレームまたは数フレーム分だけ前のフレームの撮影画像)からも特徴点を抽出したデータ(以下、特徴点空間と称する)と、抽出した特徴点とを突合することにより、ユーザの視線(位置および方向)を推定する。ここで、特徴点空間は、予め決められた利用シーンに基づいて設定された「周辺空間」に対して構築された空間である。そのため、撮影画像も当該周辺空間を撮影したものになる。そのため、特徴点空間および撮影画像の両方とも「周辺空間」に基づいて位置関係を同定していることを前提とする。 The video processing device 1 extracts feature points from the image taken by the camera 4. Furthermore, the video processing device 1 also extracts feature points from a captured image that has been captured in advance (for example, a captured image of the previous frame or several frames before the image) (hereinafter referred to as feature point space). ) and the extracted feature points to estimate the user's line of sight (position and direction). Here, the feature point space is a space constructed for a "surrounding space" set based on a predetermined usage scene. Therefore, the captured image is also a captured image of the surrounding space. Therefore, it is assumed that the positional relationships of both the feature point space and the captured image are identified based on the "surrounding space."
 映像処理装置1は、慣性センサ5から受信した慣性データに基づいてユーザの頭の動きを追跡し、上述したユーザの視線をよりリアルタイムに追従する。 The video processing device 1 tracks the movement of the user's head based on the inertial data received from the inertial sensor 5, and tracks the user's line of sight described above in more real time.
 さらに、映像処理装置1は、リアルタイムに追従しているユーザの視線を、予め作成されたARコンテンツ空間上に位置付け、そこからのARコンテンツの見え方を算出する。 Furthermore, the video processing device 1 positions the user's line of sight, which is being followed in real time, on an AR content space created in advance, and calculates how the AR content looks from there.
 そして、映像処理装置1は、出力装置3に当該算出した見え方を描画させる。 Then, the video processing device 1 causes the output device 3 to draw the calculated appearance.
 このようにして、ARシステムでは、ARコンテンツを生成し、映像処理装置1であるスマートフォンまたはARグラスに当該ARコンテンツを表示させる。しかしながら、上述したように、この方法では、カメラ4が環境を映したものであることを前提としているため、環境と移動体MBとが混在する撮影画像の場合、映像処理装置1は、ARコンテンツを正しい位置に表示させることに失敗することがある。 In this way, the AR system generates AR content and causes the video processing device 1, such as a smartphone or AR glasses, to display the AR content. However, as described above, this method assumes that the camera 4 reflects the environment, so in the case of a captured image that includes both the environment and the mobile object MB, the video processing device 1 can process the AR content. may fail to display in the correct position.
 そこで、以下では、環境と移動体とが混在する撮影画像であっても正しい位置にARコンテンツを表示させるための映像処理装置1の動作について説明する。 Therefore, the operation of the video processing device 1 for displaying AR content in the correct position even in a photographed image in which the environment and a moving object coexist will be described below.
 図3は、映像処理装置1が撮影画像の正しい位置にARコンテンツを表示させるための動作の一例を示すフローチャートである。
 映像処理装置1の制御部10がプログラム記憶部20に記憶されたプログラムを読み出して実行することにより、このフローチャートの動作が実現される。
FIG. 3 is a flowchart illustrating an example of an operation by which the video processing device 1 displays AR content at a correct position in a captured image.
The operation of this flowchart is realized by the control unit 10 of the video processing device 1 reading and executing the program stored in the program storage unit 20.
 この動作フローは、例えば、ユーザがARコンテンツの表示を望む指示を入力した、または所定の条件を満たしたため、ARコンテンツを表示する指示を制御部10が出力したことにより開始される。或いは、この動作フローは、映像処理装置1が起動し、カメラ4が撮影画像を取得した場合に開始されても良い。また、この動作において、移動体は、自転車であるとする。 This operation flow is started, for example, when the user inputs an instruction to display AR content or when a predetermined condition is satisfied, and the control unit 10 outputs an instruction to display AR content. Alternatively, this operation flow may be started when the video processing device 1 is activated and the camera 4 acquires a photographed image. Further, in this operation, it is assumed that the moving object is a bicycle.
 ステップST101で、画像取得部101は、カメラ4が撮影した撮影画像を取得する。なお、画像取得部101は、撮影画像を画像記憶部301に記憶させて良い。なお、撮影画像は、環境および移動体が含まれるものとする。ここで、環境は、上述したように、一般的な風景であって良い。そのため、環境は、移動体を除いた部分を指す。 In step ST101, the image acquisition unit 101 acquires a photographed image taken by the camera 4. Note that the image acquisition unit 101 may store the captured image in the image storage unit 301. Note that the photographed image includes the environment and a moving object. Here, the environment may be a general landscape, as described above. Therefore, the environment refers to the part excluding the moving object.
 図4は、撮影画像の一例を示す図である。
 図4の例では、ユーザが移動体である自転車に乗り、運転している際の撮影画像である。そのため、撮影画像は、移動体および環境が含まれたものである。ここで、カメラ4は、映像処理装置1であるARグラスが具備するカメラ4であり、撮影画像は、このカメラ4により撮影されたものである。
FIG. 4 is a diagram showing an example of a photographed image.
In the example of FIG. 4, the image is taken while the user is riding and driving a bicycle, which is a moving object. Therefore, the photographed image includes the moving object and the environment. Here, the camera 4 is a camera 4 included in the AR glasses that are the video processing device 1, and the photographed image is taken by this camera 4.
 また、図4の例では簡単化のため、ユーザの腕と自転車のハンドルおよび車輪が斜線部で表されている。さらに、図3の例では、撮影画像には、2つの建物Buが写っている。また、図4の例で示されるユーザは、前方を向いているとする。 Furthermore, in the example of FIG. 4, the user's arms, the bicycle handlebars, and wheels are shown with diagonal lines for the sake of simplicity. Furthermore, in the example of FIG. 3, two buildings Bu are shown in the photographed image. Further, it is assumed that the user shown in the example of FIG. 4 is facing forward.
 ステップST102で、移動体検出部102は、撮影画像から移動体MBを検出する。ここで、移動体検出部102は、一般的な方法で物体検出を行い、移動体MBを検出して良い。例えば、物体検出の方法は、非特許文献2に開示されるような物体検出の方法を使用して良い。そのため、物体検出の方法の詳細な説明はここでは省略する。 In step ST102, the moving object detection unit 102 detects the moving object MB from the captured image. Here, the moving body detection unit 102 may perform object detection using a general method to detect the moving body MB. For example, as the object detection method, an object detection method as disclosed in Non-Patent Document 2 may be used. Therefore, a detailed explanation of the object detection method will be omitted here.
 図5は、撮影画像において移動体MBが検出された際の一例を示す図である。
 図5の例では、自転車およびユーザが移動体MBとして検出される。すなわち、移動体MBとして検出される部分は、自転車のハンドルおよび車輪に加えて、ユーザの腕も含む。
FIG. 5 is a diagram illustrating an example when a moving body MB is detected in a photographed image.
In the example of FIG. 5, a bicycle and a user are detected as mobile objects MB. That is, the portion detected as the mobile body MB includes the user's arm in addition to the bicycle handle and wheels.
 ステップST103で、環境画像復元部103は、移動体MBとして検出された部分の映像を補完する。環境画像復元部103は、移動体MBとして検出された部分の映像における周辺環境の画像を補完する。すなわち、環境画像復元部103は、移動体MBとして検出された部分の周辺画像を再現する。環境画像復元部103は、一般的な方法で補完を行って良い。例えば、環境画像復元部103は、非特許文献3に開示されるような補完技術を使用して良い。そのため、補完技術の詳細な説明はここでは省略する。 In step ST103, the environment image restoration unit 103 complements the video of the portion detected as the mobile object MB. The environment image restoration unit 103 complements the image of the surrounding environment in the video of the portion detected as the mobile object MB. That is, the environmental image restoration unit 103 reproduces the peripheral image of the portion detected as the mobile object MB. The environment image restoration unit 103 may perform complementation using a general method. For example, the environmental image restoration unit 103 may use a complementary technique as disclosed in Non-Patent Document 3. Therefore, a detailed explanation of the complementary technology will be omitted here.
 図6は、移動体MBとして抽出された部分の映像を周辺画像で補完した場合の一例を示した図である。
 図6に示すように、周辺画像で移動体MBの部分の画像を補完することにより、環境画像復元部103は、移動体MBが存在しなかった場合の撮影画像を取得することができる。
FIG. 6 is a diagram showing an example of a case where a video of a portion extracted as a mobile object MB is complemented with surrounding images.
As shown in FIG. 6, by complementing the image of the mobile body MB with the surrounding image, the environmental image restoration unit 103 can obtain a photographed image that would have been taken if the mobile body MB did not exist.
 ステップST104で、視線推定部105は、ユーザの視線を推定する。最初に、特徴点抽出部104が撮影画像から特徴点を抽出する。特徴点は、例えば、2つの建物Buであって良い。 In step ST104, the line of sight estimation unit 105 estimates the user's line of sight. First, the feature point extraction unit 104 extracts feature points from the captured image. The feature points may be, for example, two buildings Bu.
 図7は、特徴点に対応するARコンテンツの設定例を示した図である。
 図7に示すように、空間記憶部302は、2つの建物Buを特徴点とした「特徴点空間」を予め設定しておき、当該空間における2つの建物Buの中間部にARコンテンツであるスマイルマークを配置した「ARコンテンツ空間」を併せて記憶しているとする。すなわち、空間記憶部302は、特徴点空間と当該特徴点空間に対応したARコンテンツ空間とを記憶している。ここで、図7では、ARコンテンツを参照符号ARCとして示してある。図7の例は一例であり、空間記憶部302は、様々な特徴点空間と共にARコンテンツ空間を記憶していて良い。
FIG. 7 is a diagram showing an example of setting AR content corresponding to feature points.
As shown in FIG. 7, the space storage unit 302 presets a "feature point space" with two buildings Bu as feature points, and places a smiley face that is AR content in the middle part of the two buildings Bu in the space. It is assumed that an "AR content space" in which marks are placed is also stored. That is, the space storage unit 302 stores a feature point space and an AR content space corresponding to the feature point space. Here, in FIG. 7, the AR content is indicated by reference symbol ARC. The example in FIG. 7 is just an example, and the space storage unit 302 may store the AR content space together with various feature point spaces.
 特徴点抽出部104は、撮影画像全体を見渡して、モノの境界(エッジ)または物体の角(コーナー)等の特定の特徴点を抽出する。ここで、モノは、物体の境界等であって良い。例えば、図7では、特定の特徴点となる建物Buの境界を抽出することが可能である。 The feature point extraction unit 104 looks over the entire photographed image and extracts specific feature points such as boundaries (edges) of objects or corners of objects. Here, the thing may be a boundary between objects. For example, in FIG. 7, it is possible to extract the boundary of the building Bu, which is a specific feature point.
 そして、視線推定部105は、特徴点抽出部104が抽出した特徴点を空間記憶部302に記憶された特徴点空間と突合することにより、視線の位置を推定する。例えば、視線推定部105は、ビジョンベースのAR技術等を用いて視線を推定して良い。ビジョンベースのAR技術は、例えば、マーカレスAR技術である、PTAM、SmartAR、Microsoft Hololens等の一般的な技術であって良い。そのため、AR技術についての詳細な説明はここでは省略する。視線推定部105は、推定した視線の位置を視線移動推定部106に出力する。 Then, the line of sight estimation unit 105 estimates the position of the line of sight by comparing the feature points extracted by the feature point extraction unit 104 with the feature point space stored in the spatial storage unit 302. For example, the line of sight estimation unit 105 may estimate the line of sight using vision-based AR technology or the like. The vision-based AR technology may be a general technology such as PTAM, SmartAR, Microsoft Hololens, etc., which are markerless AR technologies, for example. Therefore, a detailed explanation of the AR technology will be omitted here. The line-of-sight estimation unit 105 outputs the estimated line-of-sight position to the line-of-sight movement estimation unit 106.
 例えば、空間記憶部302に撮影画像に対応する画像を記憶しておき、画像同士の照合により、視線の位置を推定する方法では効率が良くない。そこで、視線推定部105は、画像そのものではなく、特徴点として抽出して単純化された特徴点と特徴点空間を突合することにより、高速に処理することが可能となる。 For example, a method of storing images corresponding to photographed images in the spatial storage unit 302 and estimating the line-of-sight position by comparing the images is not efficient. Therefore, the line-of-sight estimation unit 105 can perform high-speed processing by comparing feature points extracted and simplified as feature points with the feature point space instead of the image itself.
 ユーザが前方を向いている場合、撮影画像に写る移動体の面積が小さい。そのため、撮影画像に占める移動体についての面積が小さいことにより、環境画像復元部103は、移動体として検出された部分の補完が可能となる(画像の復元精度が高い)。その結果、視線推定部105は、視線の推定が成功することになる。 When the user is facing forward, the area of the moving object reflected in the captured image is small. Therefore, since the area occupied by the moving object in the photographed image is small, the environmental image restoration unit 103 can complement the portion detected as the moving object (image restoration accuracy is high). As a result, the line of sight estimating unit 105 successfully estimates the line of sight.
 ステップST105で、視線移動推定部106は、視線移動を推定する。最初に、センサデータ制御部107は、慣性センサ5からセンサデータを取得する。そしてセンサデータ制御部107は、取得したセンサデータからユーザの頭の動き、体の動き等を計測する。具体的には、例えば、慣性センサ5が慣性計測装置(IMU)であり、センサデータ制御部107は、慣性センサ5から加速度、角速度、地磁気等のセンサデータを取得する。そしてセンサデータ制御部107は、これらのデータに基づいてユーザの3次元の動き(例えば、ユーザの頭の動き)を計測して良い。そして、センサデータ制御部107は、計測結果を視線移動推定部106に出力する。 In step ST105, the line-of-sight movement estimation unit 106 estimates the line-of-sight movement. First, the sensor data control unit 107 acquires sensor data from the inertial sensor 5. Then, the sensor data control unit 107 measures the user's head movement, body movement, etc. from the acquired sensor data. Specifically, for example, the inertial sensor 5 is an inertial measurement unit (IMU), and the sensor data control unit 107 acquires sensor data such as acceleration, angular velocity, and geomagnetism from the inertial sensor 5. The sensor data control unit 107 may then measure the user's three-dimensional movement (for example, the user's head movement) based on these data. Then, the sensor data control unit 107 outputs the measurement result to the line of sight movement estimation unit 106.
 視線移動推定部106は、センサデータ制御部107により計測された3次元の動きを視線推定部105から受信した視線の位置を起点に移動させることにより、ユーザの頭の動きに追従した視線移動を推定する。すなわち、視線移動推定部106は、センサデータに基づいて、視線推定部105により推定された視線の位置を起点としたユーザの視線の移動を推定する。そして、視線移動推定部106は、推定した視線移動を含む視線移動情報をARコンテンツ描画部108に出力する。 The line-of-sight movement estimating unit 106 moves the three-dimensional movement measured by the sensor data control unit 107 from the position of the line-of-sight received from the line-of-sight estimation unit 105 as a starting point, thereby generating a line-of-sight movement that follows the movement of the user's head. presume. That is, the line-of-sight movement estimating unit 106 estimates the movement of the user's line of sight starting from the line-of-sight position estimated by the line-of-sight estimation unit 105 based on the sensor data. Then, the eye movement estimation unit 106 outputs eye movement information including the estimated eye movement to the AR content drawing unit 108.
 ステップST106で、ARコンテンツ描画部108は、ARコンテンツの見え方を算出する。ARコンテンツ描画部108は、視線推定部105から受信した視線移動情報に含まれる推定した視線移動、すなわち、移動先の視線を空間記憶部302に記憶された特徴点空間に対応するARコンテンツ空間に設定し、設定された空間上でのARコンテンツの見え方を算出する。 In step ST106, the AR content drawing unit 108 calculates how the AR content looks. The AR content drawing unit 108 draws the estimated line-of-sight movement included in the line-of-sight movement information received from the line-of-sight estimation unit 105, that is, the movement destination line of sight, into the AR content space corresponding to the feature point space stored in the spatial storage unit 302. and calculates how the AR content will appear in the set space.
 図8は、算出されたARコンテンツの見え方の一例を示した図である。
 図8に示すように、ARコンテンツ描画部108は、視線移動の動き情報に基づいてARコンテンツの見え方を調整し、調整したARコンテンツARCを描画するためのARコンテンツ情報を出力制御部109に出力する。
FIG. 8 is a diagram showing an example of how the calculated AR content looks.
As shown in FIG. 8, the AR content drawing unit 108 adjusts the appearance of the AR content based on the movement information of the eye movement, and sends AR content information to the output control unit 109 for drawing the adjusted AR content ARC. Output.
 ステップST107で、出力制御部109は、ARコンテンツ情報を出力する。出力制御部109は、出力装置3にARコンテンツARCを描画するように制御する。例えば、出力制御部109は、調整されたARコンテンツARCをARグラス等に表示するように制御する。 In step ST107, the output control unit 109 outputs AR content information. The output control unit 109 controls the output device 3 to draw the AR content ARC. For example, the output control unit 109 controls the adjusted AR content ARC to be displayed on AR glasses or the like.
 (第1の実施形態の作用効果) 
 第1の実施形態によれば、映像処理装置1は、カメラ4の撮影画像中に環境を映した部分と、移動体MBを映した部分とが混在する場合であっても、正確にARコンテンツARCをユーザに提示することができる。
(Operations and effects of the first embodiment)
According to the first embodiment, the video processing device 1 can accurately process AR content even when a portion showing the environment and a portion showing the mobile object MB coexist in an image taken by the camera 4. The ARC can be presented to the user.
 [第2の実施形態] 
 第2の実施形態では、例えば、自転車等の移動体に乗ったユーザが下を向いた場合、カメラ4で撮影された撮影画像の大部分を移動体が占めることになる。
[Second embodiment]
In the second embodiment, for example, when a user riding a moving object such as a bicycle looks down, the moving object will occupy most of the image taken by the camera 4.
 図9は、第2の実施形態におけるカメラ4で撮影された撮影画像の一例を示す図である。 
 図9に示すように移動体MBが撮影画像の大部分を占めるような場合、第1の実施形態での処理をしても補完処理が失敗する場合がある。
FIG. 9 is a diagram showing an example of a captured image captured by the camera 4 in the second embodiment.
As shown in FIG. 9, when the mobile object MB occupies most of the photographed image, the complementation processing may fail even if the processing in the first embodiment is performed.
 第2の実施形態では、ユーザが下を向いたりすることにより、撮影画像の大部分を移動体が占める場合であっても正確にARコンテンツをユーザに提示することを可能にする方法について説明する。 In the second embodiment, a method will be described in which it is possible to accurately present AR content to a user even when a moving object occupies most of the captured image by causing the user to look down. .
 (構成) 
 第2の実施形態における映像処理装置1のハードウェア構成は、第1の実施形態のハードウェア構成と同じで良いため、ここでの重複した説明を省略する。
(composition)
The hardware configuration of the video processing device 1 in the second embodiment may be the same as the hardware configuration in the first embodiment, so a redundant explanation here will be omitted.
 図10は、第2の実施形態における映像処理装置1のソフトウェア構成を、図1に示したハードウェア構成に関連付けて示すブロック図である。
 第2の実施形態において、制御部10は、移動体面積算出部110および路面映像解析部111と、を備える点で第1の実施形態の制御部10と異なる。
FIG. 10 is a block diagram showing the software configuration of the video processing device 1 in the second embodiment in relation to the hardware configuration shown in FIG. 1.
In the second embodiment, the control unit 10 differs from the control unit 10 in the first embodiment in that it includes a moving object area calculation unit 110 and a road image analysis unit 111.
 移動体面積算出部110は、移動体MBの面積を算出する。移動体面積算出部110は、移動体MBの面積を算出しても良いし、撮影画像に占める移動体MBの割合を算出しても良い。また、移動体面積算出部110は、算出した移動体MBの面積が閾値以上であるかどうかを判定する。 The moving body area calculation unit 110 calculates the area of the moving body MB. The moving object area calculation unit 110 may calculate the area of the moving object MB, or may calculate the ratio of the moving object MB to the captured image. Furthermore, the moving object area calculation unit 110 determines whether the calculated area of the moving object MB is equal to or larger than a threshold value.
 路面映像解析部111は、撮影画像に映る環境内の凹凸がカメラ4のシャッタースピードの間隔内で移動したことによる縦長に変形している部分を検出する。路面映像解析部111は、一般的な方法で、撮影画像の環境内に写った凹凸(例えば小石等)を検出し、当該凹凸である小石の変形を検出して良い。さらに路面映像解析部111は、その変形の度合いから移動体MBの移動速度を推定する。移動速度の推定方法の詳細は、後述する。 The road surface image analysis unit 111 detects a portion where the unevenness in the environment shown in the photographed image has been deformed into a vertically elongated shape due to movement within the interval of the shutter speed of the camera 4. The road surface image analysis unit 111 may detect irregularities (for example, pebbles, etc.) appearing in the environment of the photographed image using a general method, and detect deformation of the pebbles that are the irregularities. Furthermore, the road surface image analysis unit 111 estimates the moving speed of the mobile body MB from the degree of deformation. Details of the method for estimating the moving speed will be described later.
 また、データ記憶部30は、視線記憶部303を備える点で第1の実施形態と異なる。視線記憶部303は、視線推定部105が推定した視線についての情報を記憶して良い。なお、記憶した視線についての情報は、一定時間が経過した後、削除して良い。 Furthermore, the data storage unit 30 differs from the first embodiment in that it includes a line-of-sight storage unit 303. The line of sight storage unit 303 may store information about the line of sight estimated by the line of sight estimation unit 105. Note that the stored information regarding the line of sight may be deleted after a certain period of time has passed.
 (動作) 
 図11は、映像処理装置1が撮影画像の正しい位置にARコンテンツを表示させるための動作の一例を示すフローチャートである。
 映像処理装置1の制御部10がプログラム記憶部20に記憶されたプログラムを読み出して実行することにより、このフローチャートの動作が実現される。
(motion)
FIG. 11 is a flowchart illustrating an example of an operation by which the video processing device 1 displays AR content at a correct position in a captured image.
The operation of this flowchart is realized by the control unit 10 of the video processing device 1 reading and executing the program stored in the program storage unit 20.
 この動作フローは、例えば、ユーザがARコンテンツの表示を望む指示を入力した、または所定の条件を満たしたため、ARコンテンツを表示する指示を制御部10が出力したことにより開始される。或いは、この動作フローは、映像処理装置1が起動し、カメラ4が撮影画像を取得した場合に開始されても良い。 This operation flow is started, for example, when the user inputs an instruction to display AR content or when a predetermined condition is satisfied, and the control unit 10 outputs an instruction to display AR content. Alternatively, this operation flow may be started when the video processing device 1 is activated and the camera 4 acquires a photographed image.
 ステップST201およびステップST202は、図3を参照して説明したステップST101およびステップST102と同様であって良いため、ここでの重複した説明を省略する。なお、ステップST202で、移動体検出部102は、撮影画像および検出した移動体MBについての情報を移動体面積算出部110に出力して良い。 Step ST201 and step ST202 may be the same as step ST101 and step ST102 described with reference to FIG. 3, so duplicate explanation here will be omitted. Note that in step ST202, the moving object detection section 102 may output the captured image and information about the detected moving object MB to the moving object area calculation section 110.
 ステップST203で、移動体面積算出部110は、移動体MBの面積を算出する。 In step ST203, the moving body area calculation unit 110 calculates the area of the moving body MB.
 移動体面積算出部110は、移動体MBの面積を算出しても良いし、撮影画像に占める移動体MBの割合を算出しても良い。 The moving object area calculation unit 110 may calculate the area of the moving object MB, or may calculate the proportion of the moving object MB in the captured image.
 ステップST204で、移動体面積算出部110は、算出した移動体MBの面積が閾値以上であるかどうかを判定する。算出した移動体MBの面積が所定の閾値以上でないと判定した、すなわち移動体MBの面積が撮影画像の大部分を占めていない場合、処理は、ステップST205に進む。一方、算出した面積が所定の閾値以上であると判定した、すなわち、移動体MBの面積が撮影画像の大部分を占めている場合、処理は、ステップST207に進む。なお、移動体面積算出部110は、割合を算出した場合、移動体MBの割合が閾値以上であるかどうか判定して良い。 ステップST205およびステップST206は、図3を参照して説明したステップST103およびステップST104と同様であって良いため、ここでの重複した説明を省略する。ただし、視線推定部105は、推定した視線推定についての情報を時刻情報と共に視線記憶部303に記憶させておいてよい。 In step ST204, the moving object area calculation unit 110 determines whether the calculated area of the moving object MB is greater than or equal to a threshold value. If it is determined that the calculated area of the moving body MB is not equal to or larger than the predetermined threshold, that is, if the area of the moving body MB does not occupy most of the captured image, the process proceeds to step ST205. On the other hand, if it is determined that the calculated area is equal to or greater than the predetermined threshold, that is, if the area of the moving body MB occupies most of the captured image, the process proceeds to step ST207. Note that when calculating the ratio, the moving object area calculation unit 110 may determine whether the ratio of the moving object MB is equal to or greater than a threshold value. Step ST205 and step ST206 may be the same as step ST103 and step ST104 described with reference to FIG. 3, so a duplicate description here will be omitted. However, the line-of-sight estimating unit 105 may store information about the estimated line-of-sight estimation in the line-of-sight storage unit 303 together with time information.
 例えば、図7に示すように、ユーザが下を向いている場合、撮影画像の大部分を移動体が占める。このような場合に、第1の実施形態と同様に、環境画像復元部103が移動体として検出された部分の補完を行おうとしてもその精度が劣化し、復元が正しくできない。その結果、視線推定部104は、視線の推定に失敗する。そこで、処理は、以下で説明するように、ユーザが下を向いた場合でも視線の推定を行える手法でユーザの視線を推定することになる。 For example, as shown in FIG. 7, when the user is looking down, the moving object occupies most of the captured image. In such a case, as in the first embodiment, even if the environmental image restoration unit 103 tries to complement the portion detected as a moving object, the accuracy deteriorates and restoration cannot be performed correctly. As a result, the line of sight estimation unit 104 fails to estimate the line of sight. Therefore, as described below, the process estimates the user's line of sight using a method that allows estimation of the line of sight even when the user looks down.
 ステップST207で、路面映像解析部111は、撮影画像に映る環境内の凹凸がカメラ4のシャッタースピードの間隔内で移動したことによる縦長に変形している部分を検出する。路面映像解析部111は、一般的な方法で、撮影画像の環境内に写った凹凸(例えば小石等)を検出し、当該凹凸である小石の変形を検出して良い。例えば、凹凸(小石)は、(凹凸)に外接する長方形として検出される。そこで、路面映像解析部111は、この長方形の縦と横の長さの比から変形の度合いを推定する。そして、路面映像解析部111は、その変形の度合いから移動体MBの移動速度を推定する。路面映像解析部111は、例えば、ブラーの強度、カメラ4の取付け角度等に基づいて、撮影画像に映っている移動体MBの速度を推定する。ここで、移動体MBの速度の推定方法は、一般的な技術を用いてよい。例えば、移動体MBの速度推定の方法は、非特許文献4に開示されるような移動体MBの速度推定の方法を使用して良い。そのため、移動体MBの速度推定の方法の詳細な説明はここでは省略する。 In step ST207, the road image analysis unit 111 detects a portion where the unevenness in the environment shown in the photographed image has been deformed into a vertically elongated shape due to movement within the interval of the shutter speed of the camera 4. The road surface image analysis unit 111 may detect irregularities (for example, pebbles, etc.) appearing in the environment of the photographed image using a general method, and detect deformation of the pebbles that are the irregularities. For example, irregularities (pebbles) are detected as a rectangle circumscribing the irregularities. Therefore, the road image analysis unit 111 estimates the degree of deformation from the ratio of the length and width of this rectangle. Then, the road image analysis unit 111 estimates the moving speed of the mobile body MB from the degree of deformation. The road surface video analysis unit 111 estimates the speed of the moving object MB shown in the photographed image based on, for example, the intensity of blur, the mounting angle of the camera 4, and the like. Here, as a method for estimating the speed of the mobile body MB, a general technique may be used. For example, as the method for estimating the speed of the mobile body MB, a method for estimating the speed of the mobile body MB as disclosed in Non-Patent Document 4 may be used. Therefore, a detailed explanation of the method for estimating the speed of the mobile body MB will be omitted here.
 図12は、撮影画像における路面上の凹凸を検出した一例を示した図である。
 図12では、路面上の凹凸を参照符号CCで示してある。図12に示すように、路面映像解析部111は、撮影画像から路面上の小石等の凹凸を検出する。
FIG. 12 is a diagram showing an example of detecting unevenness on a road surface in a photographed image.
In FIG. 12, unevenness on the road surface is indicated by reference symbol CC. As shown in FIG. 12, the road image analysis unit 111 detects irregularities such as pebbles on the road surface from the photographed image.
 図13は、撮影画像における路面上の凹凸に対してブラーが掛かっている一例を示した図である。
 図13に示すように、移動体MBの速度が速くなればなるほど検出した小石に対してブラーが掛かるようになる。路面映像解析部111は、このブラーの度合いを用いて移動体MBの速度を推定して良い。
FIG. 13 is a diagram showing an example in which unevenness on a road surface in a photographed image is blurred.
As shown in FIG. 13, as the speed of the moving body MB increases, the detected pebbles become blurred. The road surface image analysis unit 111 may estimate the speed of the moving body MB using the degree of blur.
 ステップST208で、視線推定部105は、視線を推定する。視線推定部105は、現在の時刻から前回記憶した視線推定についての情報を視線記憶部303から取得し、当該視線推定の結果および移動速度に基づいて、現在の視線位置を推定する。 In step ST208, the line of sight estimation unit 105 estimates the line of sight. The line-of-sight estimation unit 105 acquires information about the previously stored line-of-sight estimation from the current time from the line-of-sight storage unit 303, and estimates the current line-of-sight position based on the result of the line-of-sight estimation and the movement speed.
 ステップST209~ステップST211は、図3を参照して説明したステップST105~ステップST107と同様であって良いため、ここでの重複した説明を省略する。 Steps ST209 to ST211 may be the same as steps ST105 to ST107 described with reference to FIG. 3, so a duplicate explanation here will be omitted.
 例えば、ARコンテンツ空間は、現実世界の経度および緯度のような座標系と紐づけられている。そのため、視線推定の方法が撮影画像中の凹凸の歪みを利用する方法になった場合であっても、速度と時間で算出される移動距離と、経度緯度の座標系におけるARコンテンツ空間の2点間の距離とで相互に紐付けおよびマッピングが可能である。そのため、ステップST209~ステップST211による処理で、映像処理装置1は、ARコンテンツ空間を正しく認識することが可能となり、ARコンテンツARCを正確にユーザに提示することができる。 For example, the AR content space is linked to a coordinate system such as longitude and latitude in the real world. Therefore, even if the line of sight estimation method is a method that utilizes the distortion of unevenness in the captured image, there are two points in the AR content space: the moving distance calculated from speed and time, and the longitude/latitude coordinate system. It is possible to link and map each other based on the distance between them. Therefore, through the processing in steps ST209 to ST211, the video processing device 1 can correctly recognize the AR content space and can accurately present the AR content ARC to the user.
 (第2の実施形態の作用効果) 
 第2の実施形態によれば、映像処理装置1は、カメラ4の撮影画像において、環境を映した部分よりも移動体MBを映した部分が大部分を占める場合であっても、正確にARコンテンツARCをユーザに提示することができる。
(Operations and effects of the second embodiment)
According to the second embodiment, the video processing device 1 can accurately perform AR processing even when the portion of the image captured by the camera 4 is dominated by the portion of the mobile object MB rather than the portion of the environment. The content ARC can be presented to the user.
 [他の実施形態]
 上記の実施形態では、映像処理装置1が具備するカメラ4で撮影された撮影画像を用いる例を説明したが、撮影画像は、映像処理装置1が具備したカメラ4であることに限られない。例えば、映像処理装置1に接続された独立したカメラ4であって良い。ただし、カメラ4は、ユーザの視線を推定することができる撮影画像が撮影可能な場所(例えば、ユーザの頭の上等)に設置されているとする。
[Other embodiments]
In the above embodiment, an example has been described in which a photographed image photographed by the camera 4 included in the video processing device 1 is used, but the photographed image is not limited to the camera 4 included in the video processing device 1. For example, it may be an independent camera 4 connected to the video processing device 1. However, it is assumed that the camera 4 is installed at a location (for example, above the user's head) where a captured image from which the user's line of sight can be estimated can be captured.
 また、前記実施形態に記載した手法は、計算機(コンピュータ)に実行させることができるプログラム(ソフトウェア手段)として、例えば磁気ディスク(フロッピー(登録商標)ディスク、ハードディスク等)、光ディスク(CD-ROM、DVD、MO等)、半導体メモリ(ROM、RAM、フラッシュメモリ等)等の記憶媒体に格納し、また通信媒体により伝送して頒布することもできる。なお、媒体側に格納されるプログラムには、計算機に実行させるソフトウェア手段(実行プログラムのみならずテーブル、データ構造も含む)を計算機内に構成させる設定プログラムをも含む。本装置を実現する計算機は、記憶媒体に記憶されたプログラムを読み込み、また場合により設定プログラムによりソフトウェア手段を構築し、このソフトウェア手段によって動作が制御されることにより上述した処理を実行する。なお、本明細書で言う記憶媒体は、頒布用に限らず、計算機内部或いはネットワークを介して接続される機器に設けられた磁気ディスク、半導体メモリ等の記憶媒体を含むものである。 Furthermore, the method described in the above embodiments can be applied to, for example, magnetic disks (floppy (registered trademark) disks, hard disks, etc.), optical disks (CD-ROMs, DVDs, etc.) as programs (software means) that can be executed by a computer. , MO, etc.), semiconductor memory (ROM, RAM, flash memory, etc.), and can also be transmitted and distributed via a communication medium. Note that the programs stored on the medium side also include a setting program for configuring software means (including not only execution programs but also tables and data structures) in the computer to be executed by the computer. A computer that realizes this device reads a program stored in a storage medium, and if necessary, constructs software means using a setting program, and executes the above-described processing by controlling the operation of the software means. Note that the storage medium referred to in this specification is not limited to those for distribution, and includes storage media such as magnetic disks and semiconductor memories provided inside computers or devices connected via a network.
 要するに、この発明は上記実施形態に限定されるものではなく、実施段階ではその要旨を逸脱しない範囲で種々に変形することが可能である。また、各実施形態は可能な限り適宜組み合わせて実施してもよく、その場合組み合わせた効果が得られる。さらに、上記実施形態には種々の段階の発明が含まれており、開示される複数の構成要件における適当な組み合わせにより種々の発明が抽出され得る。 In short, the present invention is not limited to the above-described embodiments, and various modifications can be made at the implementation stage without departing from the spirit thereof. Moreover, each embodiment may be implemented by appropriately combining them as much as possible, and in that case, the combined effects can be obtained. Further, the embodiments described above include inventions at various stages, and various inventions can be extracted by appropriately combining the plurality of disclosed constituent elements.
 1…映像処理装置
 2…入力装置
 3…出力装置
 4…カメラ
 5…慣性センサ
 10…制御部
 101…画像取得部
 102…移動体検出部
 103…環境画像復元部
 104…特徴点抽出部
 105…視線推定部
 106…視線移動推定部
 107…センサデータ制御部
 108…ARコンテンツ描画部
 109…出力制御部
 110…移動体面積算出部
 111…路面映像解析部
 20…プログラム記憶部
 30…データ記憶部
 301…画像記憶部
 302…空間記憶部
 303…視線記憶部
 40…通信インタフェース
 50…入出力インタフェース
 MB…移動体
 Bu…建物
1...Video processing device 2...Input device 3...Output device 4...Camera 5...Inertial sensor 10...Control unit 101...Image acquisition unit 102...Moving object detection unit 103...Environmental image restoration unit 104...Feature point extraction unit 105...Line of sight Estimation unit 106... Gaze movement estimation unit 107... Sensor data control unit 108... AR content drawing unit 109... Output control unit 110... Moving object area calculation unit 111... Road image analysis unit 20... Program storage unit 30... Data storage unit 301... Image storage unit 302... Spatial storage unit 303... Line of sight storage unit 40... Communication interface 50... Input/output interface MB... Mobile object Bu... Building

Claims (8)

  1.  ユーザが装着する映像処理装置であって、
     カメラで撮影した、前記ユーザが乗っている移動体と環境とを含む撮影画像を取得する画像取得部と、
     前記撮影画像から前記移動体を検出する移動体検出部と、
     前記検出された移動体の面積が所定の閾値より小さい場合、前記検出された移動体の部分における環境の画像を補完する環境画像復元部と、
     前記撮影画像に基づいて前記ユーザの視線を推定する視線推定部と、
     前記映像処理装置が備えるセンサからセンサデータを取得し、前記センサデータに基づいて、前記推定された視線を起点とした前記ユーザの視線の移動を検出する視線移動検出部と、
     前記推定された視線の移動に基づいて、ARコンテンツの見え方を算出する描画部と、
     前記算出されたARコンテンツを表示するように制御する出力制御部と、
     を備える映像処理装置。
    A video processing device worn by a user,
    an image acquisition unit that acquires an image captured by a camera that includes the moving object on which the user is riding and the environment;
    a moving object detection unit that detects the moving object from the captured image;
    an environment image restoration unit that complements an image of the environment in a portion of the detected moving body when the area of the detected moving body is smaller than a predetermined threshold;
    a line-of-sight estimation unit that estimates the user's line-of-sight based on the captured image;
    a line-of-sight movement detection unit that acquires sensor data from a sensor included in the video processing device, and detects movement of the user's line of sight starting from the estimated line-of-sight based on the sensor data;
    a drawing unit that calculates how the AR content looks based on the estimated movement of the line of sight;
    an output control unit that controls to display the calculated AR content;
    An image processing device comprising:
  2.  前記検出された移動体の面積を算出し、前記面積が所定の閾値を超えるかどうかを判定する移動体面積算出部と、
     前記面積が所定の閾値を超える場合、前記撮影画像の環境内の凹凸を検出し、前記凹凸の変形を検出し、前記変形の度合いに基づいて前記移動体の速度を推定する映像解析部をさらに備える、請求項1に記載の映像処理装置。
    a moving object area calculation unit that calculates the area of the detected moving object and determines whether the area exceeds a predetermined threshold;
    If the area exceeds a predetermined threshold, further comprising a video analysis unit that detects unevenness in the environment of the photographed image, detects deformation of the unevenness, and estimates the speed of the moving object based on the degree of the deformation. The video processing device according to claim 1.
  3.  前記視線推定部は、前記移動体の速度および前回の視線推定に基づいてユーザの現在の視線を推定する、請求項2に記載の映像処理装置。 The video processing device according to claim 2, wherein the line-of-sight estimating unit estimates the user's current line-of-sight based on the speed of the moving object and the previous line-of-sight estimation.
  4.  前記撮影画像内の環境内にある特徴点を抽出する特徴点抽出部と、
     特徴点空間を記憶する記憶部と、
     をさらに備え、前記視線推定部は、前記抽出された特徴点および前記特徴点空間に基づいて視線を推定する、請求項1に記載の映像処理装置。
    a feature point extraction unit that extracts feature points in the environment within the photographed image;
    a storage unit that stores the feature point space;
    The video processing device according to claim 1, further comprising: the line of sight estimating unit estimates the line of sight based on the extracted feature points and the feature point space.
  5.  前記記憶部は、前記特徴点空間に対応したARコンテンツ空間をさらに記憶し、
     前記描画部は、前記視線の移動先を前記ARコンテンツ空間に設定し、前記設定された空間上で前記ARコンテンツの見え方を算出する、請求項4に記載の映像処理装置。
    The storage unit further stores an AR content space corresponding to the feature point space,
    The video processing device according to claim 4, wherein the drawing unit sets the destination of the line of sight to the AR content space, and calculates how the AR content looks in the set space.
  6.  特徴点抽出部は、前記撮影画像内のモノのエッジまたはコーナーにあるものを特徴点として抽出する、請求項4に記載の映像処理装置。 The video processing device according to claim 4, wherein the feature point extraction unit extracts features at edges or corners of objects in the photographed image.
  7.  ユーザが装着する映像処理装置のプロセッサが実行する映像処理方法であって、
     カメラで撮影した、前記ユーザが乗っている移動体と環境とを含む撮影画像を取得することと、
     前記撮影画像から前記移動体を検出することと、
     前記検出された移動体の面積が所定の閾値より小さい場合、前記検出された移動体の部分における環境の画像を補完することと、
     前記撮影画像に基づいて前記ユーザの視線を推定することと、
     前記映像処理装置が備えるセンサからセンサデータを取得することと、
     前記センサデータに基づいて、前記推定された視線を起点とした前記ユーザの視線の移動を検出することと、
     前記推定された視線の移動に基づいて、ARコンテンツの見え方を算出することと、
     前記算出されたARコンテンツを表示するように制御することと、
     を備える映像処理方法。
    A video processing method executed by a processor of a video processing device worn by a user, the method comprising:
    Obtaining a captured image captured by a camera that includes the moving object on which the user is riding and the environment;
    Detecting the moving object from the captured image;
    When the area of the detected moving body is smaller than a predetermined threshold, complementing an image of the environment in a portion of the detected moving body;
    Estimating the user's line of sight based on the captured image;
    acquiring sensor data from a sensor included in the video processing device;
    Detecting a movement of the user's line of sight starting from the estimated line of sight based on the sensor data;
    Calculating how the AR content looks based on the estimated movement of the line of sight;
    Controlling to display the calculated AR content;
    A video processing method comprising:
  8.  ユーザが装着する映像処理装置のプロセッサによって実行させるための命令を備える映像処理プログラムであって、前記命令は、
     カメラで撮影した、前記ユーザが乗っている移動体と環境とを含む撮影画像を取得することと、
     前記撮影画像から前記移動体を検出することと、
     前記検出された移動体の面積が所定の閾値より小さい場合、前記検出された移動体の部分における環境の画像を補完することと、
     前記撮影画像に基づいて前記ユーザの視線を推定することと、
     前記映像処理装置が備えるセンサからセンサデータを取得することと、
     前記センサデータに基づいて、前記推定された視線を起点とした前記ユーザの視線の移動を検出することと、
     前記推定された視線の移動に基づいて、ARコンテンツの見え方を算出することと、
     前記算出されたARコンテンツを表示するように制御することと、
     を備える、映像処理プログラム。
    A video processing program comprising instructions to be executed by a processor of a video processing device worn by a user, the instructions comprising:
    Obtaining a captured image captured by a camera that includes the moving object on which the user is riding and the environment;
    Detecting the moving object from the captured image;
    When the area of the detected moving body is smaller than a predetermined threshold, complementing an image of the environment in a portion of the detected moving body;
    Estimating the user's line of sight based on the captured image;
    acquiring sensor data from a sensor included in the video processing device;
    Detecting a movement of the user's line of sight starting from the estimated line of sight based on the sensor data;
    Calculating how the AR content looks based on the estimated movement of the line of sight;
    Controlling to display the calculated AR content;
    A video processing program with
PCT/JP2022/031904 2022-08-24 2022-08-24 Video processing device, video processing method, and video processing program WO2024042644A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/JP2022/031904 WO2024042644A1 (en) 2022-08-24 2022-08-24 Video processing device, video processing method, and video processing program

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2022/031904 WO2024042644A1 (en) 2022-08-24 2022-08-24 Video processing device, video processing method, and video processing program

Publications (1)

Publication Number Publication Date
WO2024042644A1 true WO2024042644A1 (en) 2024-02-29

Family

ID=90012833

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2022/031904 WO2024042644A1 (en) 2022-08-24 2022-08-24 Video processing device, video processing method, and video processing program

Country Status (1)

Country Link
WO (1) WO2024042644A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018207426A1 (en) * 2017-05-09 2018-11-15 ソニー株式会社 Information processing device, information processing method, and program
WO2019087658A1 (en) * 2017-11-01 2019-05-09 ソニー株式会社 Information processing device, information processing method, and program
JP2019128824A (en) * 2018-01-25 2019-08-01 キヤノン株式会社 Image processing apparatus and image processing method
JP2022072598A (en) * 2020-10-30 2022-05-17 株式会社小松製作所 Work vehicle display system and work vehicle display method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018207426A1 (en) * 2017-05-09 2018-11-15 ソニー株式会社 Information processing device, information processing method, and program
WO2019087658A1 (en) * 2017-11-01 2019-05-09 ソニー株式会社 Information processing device, information processing method, and program
JP2019128824A (en) * 2018-01-25 2019-08-01 キヤノン株式会社 Image processing apparatus and image processing method
JP2022072598A (en) * 2020-10-30 2022-05-17 株式会社小松製作所 Work vehicle display system and work vehicle display method

Similar Documents

Publication Publication Date Title
US10499002B2 (en) Information processing apparatus and information processing method
US10740975B2 (en) Mobile augmented reality system
US10677596B2 (en) Image processing device, image processing method, and program
JP6491517B2 (en) Image recognition AR device, posture estimation device, and posture tracking device
US9270891B2 (en) Estimation of panoramic camera orientation relative to a vehicle coordinate frame
KR101749017B1 (en) Speed-up template matching using peripheral information
JP6100380B2 (en) Image processing method used for vision-based positioning, particularly for apparatus
CN110262706B (en) Information processing apparatus, information processing method, and recording medium
US9842399B2 (en) Image processing device and image processing method
US10545031B2 (en) Portable terminal device, recording medium, and correction method
CN110794955B (en) Positioning tracking method, device, terminal equipment and computer readable storage medium
CN103492899A (en) Online reference patch generation and pose estimation for augmented reality
KR101923177B1 (en) Appratus and method for providing augmented reality information based on user
US20190088027A1 (en) Method for developing augmented reality experiences in low computer power systems and devices
KR101253644B1 (en) Apparatus and method for displaying augmented reality content using geographic information
JP2018503195A (en) Object detection method and object detection apparatus
CN108804161B (en) Application initialization method, device, terminal and storage medium
WO2024042644A1 (en) Video processing device, video processing method, and video processing program
CN111553342B (en) Visual positioning method, visual positioning device, computer equipment and storage medium
WO2024042645A1 (en) Video processing device, video processing method, and video processing program
JP2015005220A (en) Information display device and information display method
JP6393000B2 (en) Hypothetical line mapping and validation for 3D maps
JP7198966B2 (en) Positioning system
KR102339710B1 (en) Method and electronic device for displaying video information
KR102633950B1 (en) Method and electronic device for measuring velocity of object using cctv video

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22956472

Country of ref document: EP

Kind code of ref document: A1