WO2015072091A1 - Image processing device, image processing method, and program storage medium - Google Patents

Image processing device, image processing method, and program storage medium Download PDF

Info

Publication number
WO2015072091A1
WO2015072091A1 PCT/JP2014/005387 JP2014005387W WO2015072091A1 WO 2015072091 A1 WO2015072091 A1 WO 2015072091A1 JP 2014005387 W JP2014005387 W JP 2014005387W WO 2015072091 A1 WO2015072091 A1 WO 2015072091A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
background
content
unit
marker
Prior art date
Application number
PCT/JP2014/005387
Other languages
French (fr)
Japanese (ja)
Inventor
克幸 永井
幸司 森下
尚志 野田
英良 松嶋
岩元 浩太
Original Assignee
Necソリューションイノベータ株式会社
日本電気株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Necソリューションイノベータ株式会社, 日本電気株式会社 filed Critical Necソリューションイノベータ株式会社
Publication of WO2015072091A1 publication Critical patent/WO2015072091A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality

Definitions

  • the present invention relates to augmented reality (AR) image processing technology.
  • AR augmented reality
  • Non-Patent Document 1 proposes a method of generating a three-dimensional content to be superimposed on the real world with an image obtained by actually capturing a real object existing in the real world.
  • CG computer graphic
  • AR image a composite image in which 3D content and 2D data are superimposed on the real world
  • an image of the real world on which 3D content or 2D content is superimposed may be described as a background image.
  • AR content and 2D content superimposed on the background image may be referred to as AR content.
  • Non-Patent Document 2 proposes a remote control system for a mobile robot using AR technology.
  • an image in which a three-dimensional model of a robot is superimposed on an extended free viewpoint image using AR technology is presented to the operator via the HMD.
  • This extended free viewpoint image is generated based on the omnidirectional image group and distance image group acquired at a plurality of points by the mobile robot, and the position and orientation of the robot at the time of acquisition.
  • Patent Document 1 as a method for generating three-dimensional content to be superimposed, a converted image for three-dimensional display of a target object can be obtained with a single camera without using a dedicated device such as an acceleration sensor.
  • a generation method has been proposed.
  • the conversion target image is geometrically transformed so that the specified image coordinates match by allowing the user to specify a plurality of image coordinates constituting the rotation axis of the camera for a plurality of images including the target object.
  • Yusuke Uno “Presentation of Virtual Real Objects Based on View-Dependent Texture Geometry for Augmented Reality”, [online], [Search September 26, 2014], Internet ⁇ URL: http: //yokoya.naist .jp / paper / datas / 1295 / mthesis_uno_final.pdf> Yuko Ueda et al. “Remote Mobile Robot Operation Interface Using Extended Free Viewpoint Image Generation”, IEICE Technical Report, vol.112, no.386, pp.217-222, January 2013
  • a real-time video obtained from an imaging device in real time is used as a background image on which 3D content or 2D content is superimposed.
  • the user can visually recognize an AR image in which AR content exists in the real world captured by the imaging apparatus in real time.
  • the real world where the AR content is to be superimposed is limited to the location of the imaging device at that time.
  • a user can view an AR image in which AR content is superimposed on his / her room with his / her camera-equipped mobile terminal only when the user is in his / her room with the mobile terminal. .
  • the user cannot see the AR image in a place other than the user's room such as in a train.
  • the present invention has been made in view of such circumstances.
  • One of the objects of the present invention is to provide a technique for presenting an AR image in which AR content is superimposed on a desired background image without restrictions on place and time.
  • the first aspect of the present invention relates to an image processing apparatus.
  • An image processing apparatus is received by an operation receiving unit from an operation receiving unit that receives an image file selection operation and a background holding unit that stores an image file that stores a background image in which a marker is captured.
  • Background acquisition means for acquiring an image of the image file selected by the selection operation as a background image
  • marker detection means for detecting a marker from the background image acquired by the background acquisition means
  • a reference calculation means for calculating a background shooting position and a background shooting direction, and a position in the three-dimensional coordinate space where the AR content to be displayed is virtually arranged is determined.
  • the second aspect of the present invention relates to an image processing method.
  • the image processing method according to the second aspect of the present invention receives an image file selection operation, and is selected by the received selection operation from a background holding unit that holds an image file that stores a background image in which a marker is captured.
  • An image of an image file is acquired as a background image, a marker is detected from the acquired background image, a three-dimensional coordinate space is set based on the actual information and image information regarding the detected marker, and the reality regarding the detected marker
  • the background photographing position and the background photographing direction are calculated, the position on the three-dimensional coordinate space where the AR content to be displayed is virtually arranged is determined, and the determined position in the three-dimensional coordinate space
  • Display on the display means a composite image in which the visual image from the background shooting position and the background shooting direction of the AR content arranged on the background image is superimposed on the background image.
  • the third aspect of the present invention relates to a program.
  • a program comprising: an operation receiving unit that receives a selection operation of an image file; and a background holding unit that holds an image file that stores a background image in which a marker is captured.
  • a background acquisition unit that acquires an image of an image file selected by an accepted selection operation as a background image
  • a marker detection unit that detects a marker from the background image acquired by the background acquisition unit, and the detected marker
  • the reference calculation means for setting the three-dimensional coordinate space based on the real information and the image information and calculating the background photographing position and the background photographing direction and the AR (Augmented Reality) content to be displayed are virtually arranged.
  • a position determining means for determining a position in the three-dimensional coordinate space and a position determining unit arranged at the determined position in the three-dimensional coordinate space.
  • Display control means for viewing images displayed on the display means a composite image superimposed on the background image from the background photographing position and the background photographing direction of the AR content is to operate with.
  • the third aspect of the present invention is also realized by a computer-readable recording medium that records the program.
  • This recording medium includes a non-transitory tangible medium.
  • FIG. 1 is a diagram conceptually illustrating a processing configuration example of an image processing apparatus according to the first embodiment of the present invention.
  • FIG. 2 is a flowchart showing an operation example of the image processing apparatus according to the first embodiment of the present invention.
  • FIG. 3 is a diagram conceptually illustrating a hardware configuration example of an image processing apparatus according to the second embodiment of the present invention.
  • FIG. 4 is a diagram conceptually illustrating a processing configuration example of the image processing apparatus according to the second embodiment of the present invention.
  • FIG. 5 is a diagram conceptually illustrating an example of a plurality of content images held in the content holding unit and associated with various content shooting directions.
  • FIG. 6 is a flowchart showing an operation example of the image processing apparatus according to the second embodiment of the present invention.
  • FIG. 1 is a diagram conceptually illustrating a processing configuration example of an image processing apparatus according to the first embodiment of the present invention.
  • FIG. 2 is a flowchart showing an operation example of the image processing apparatus according to the first embodiment of the
  • FIG. 7 is a diagram conceptually illustrating a hardware configuration example of an image processing apparatus according to the third embodiment of the present invention.
  • FIG. 8 is a flowchart showing a part of an operation example of the image processing apparatus according to the third embodiment of the present invention.
  • FIG. 9 is a diagram conceptually illustrating a processing configuration example of the image processing apparatus according to the third modification of the present invention.
  • FIG. 10A is a diagram illustrating an example of a background image stored in an image file.
  • FIG. 10B is a diagram illustrating an example of a background image stored in an image file.
  • FIG. 10C is a diagram illustrating an example of a background image stored in an image file.
  • FIG. 10A is a diagram illustrating an example of a background image stored in an image file.
  • FIG. 10B is a diagram illustrating an example of a background image stored in an image file.
  • FIG. 10C is a diagram illustrating an example of a background image stored in an image file.
  • FIG. 11A is a diagram showing an example of a composite image (AR image) in which a content image is superimposed on the background image shown in FIG. 10A.
  • FIG. 11B is a diagram showing an example of a composite image (AR image) in which a content image is superimposed on the background image shown in FIG. 10B.
  • FIG. 11C is a diagram illustrating an example of a composite image (AR image) in which a content image is superimposed on the background image illustrated in FIG. 10C.
  • FIG. 12A is a diagram illustrating movement of AR content on an AR image.
  • FIG. 12B is a diagram illustrating movement of AR content on an AR image.
  • FIG. 12C is a diagram illustrating movement of AR content on an AR image.
  • FIG. 1 is a diagram conceptually illustrating a processing configuration example of an image processing apparatus according to the first embodiment of the present invention.
  • the image processing apparatus 100 includes an operation reception unit 101, a background acquisition unit 102, a marker detection unit 103, a reference calculation unit 104, a position determination unit 105, a display control unit 106, and the like.
  • the image processing apparatus 100 has the same hardware configuration as the image processing apparatus 10 in a detailed embodiment described later. Similar to the image processing apparatus 10, the image processing apparatus 100 is controlled by a program, thereby realizing the above-described processing units.
  • the image processing apparatus 100 may include a background holding unit 107 and a display unit 108.
  • the display unit 108 is a device that displays a screen generated based on drawing data, such as an LCD (Liquid Crystal Display) or a CRT (Cathode Ray Tube) display.
  • LCD Liquid Crystal Display
  • CRT Cathode Ray Tube
  • the operation reception unit 101 receives an image file selection operation.
  • the background acquisition unit 102 acquires, as a background image, the image of the image file selected by the selection operation received by the operation reception unit 101 from the background holding unit 107 that stores the image file that stores the background image in which the marker is reflected.
  • moving image data or still image data is stored in the image file.
  • the background acquisition unit 102 acquires a background image as still image information from moving image data or still image data.
  • the marker detection unit 103 detects a marker from the background image acquired by the background acquisition unit 102.
  • the marker is an image drawn on an object arranged in the real world or an object arranged in the real world.
  • the marker is called an AR marker or the like.
  • the marker detection unit 103 holds in advance marker actual information (such as shape information). And the marker detection part 103 detects a marker from a background image based on this real information.
  • the reference calculation unit 104 sets a three-dimensional coordinate space based on the detected real information and image information about the marker, and calculates the background shooting position and the background shooting direction.
  • the actual information related to the marker means known information in the real world of the marker.
  • the actual information regarding the marker indicates, for example, the actual shape of the marker.
  • the image information regarding a marker means the information in the background image of a marker.
  • the image information regarding the marker indicates the shape of the marker in the background image.
  • the reference calculation unit 104 sets a three-dimensional coordinate space at a position corresponding to the reference point recognized from the detected marker.
  • the position where the three-dimensional coordinate space is set is not particularly limited.
  • the reference calculation unit 104 calculates the background shooting position and the background shooting direction by comparing the actual information about the marker and the image information.
  • the background shooting position and the background shooting direction mean the position and orientation (posture) of the imaging apparatus that shot the background image of the image file.
  • the position determination unit 105 determines the position in the three-dimensional coordinate space where the AR content to be displayed is virtually arranged.
  • the position determination unit 105 can determine this position arbitrarily or fixedly.
  • the AR content to be displayed may be a three-dimensional model proposed in Non-Patent Documents 1 and 2 above.
  • the AR content to be displayed may be three-dimensional content that is simulated by switching and displaying a plurality of content images obtained by photographing a real object from a plurality of directions.
  • the AR content to be displayed may be planar two-dimensional content.
  • the display control unit 106 displays a composite image in which a visual image from the background shooting position and the background shooting direction of the AR content arranged at the determined position in the three-dimensional coordinate space is superimposed on the background image. 108 is displayed.
  • the display control unit 106 obtains the visual image using a known method based on the information of the three-dimensional model or the two-dimensional content. Can do.
  • a method for acquiring the visual image when the AR content is a three-dimensional content that is simulated by switching and displaying a plurality of content images obtained by photographing a real object from a plurality of directions is described in detail. Will be described later.
  • FIG. 2 is a flowchart showing an operation example of the image processing apparatus 100 in the present embodiment.
  • the image processing method in the present embodiment is executed by at least one computer such as an image processing apparatus 100 as shown in FIG.
  • the computer may have a hardware configuration similar to that of the image processing apparatus 10 described later.
  • the image processing method includes processing steps from step S21 to step S27.
  • step S21 the computer accepts an image file selection operation.
  • step S22 the computer acquires the image of the image file selected by the selection operation accepted in step S21 as the background image from the background holding unit 107 that holds the image files in which the markers each store the background image.
  • the marker and the image file are as described above.
  • the content of step S22 is the same as the processing content of the background acquisition unit 102.
  • step S23 the computer detects a marker from the background image acquired in step S22.
  • the content of step S23 is the same as the processing content of the marker detection unit 103.
  • step S24 the computer sets a three-dimensional coordinate space based on the real information and image information regarding the marker detected in step S23.
  • step S25 the computer calculates a background shooting position and a background shooting direction based on the actual information and the image information regarding the marker detected in step S23.
  • the contents of steps S24 and S25 are the same as the contents of processing of the reference calculation unit 104.
  • step S26 the computer determines a position on the three-dimensional coordinate space where the AR content to be displayed is virtually arranged.
  • the content of step S26 is the same as the processing content of the position determination unit 105.
  • step S27 the computer causes the display unit 108 to display a composite image in which the visual image of the AR content to be arranged is superimposed on the background image.
  • the position where the AR content is arranged is the position determined in step S26 in the three-dimensional coordinate space set in step S24.
  • the visual image of the AR content is a visual image of the AR content arranged at the above-described position from the background shooting position and the background shooting direction calculated in step S25.
  • the background image is the background image acquired in step S22.
  • the content of step S27 is the same as the processing content of the display control unit 106.
  • At least one computer that is the execution subject of the image processing method may include the display unit 108 and the background holding unit 107.
  • Another computer (not shown) connected to be communicable with the at least one computer may include the display unit 108 and the background holding unit 107.
  • each embodiment of the present invention may be realized by a program that causes at least one computer to execute the above-described image processing method.
  • Each embodiment of the present invention may be realized by a computer-readable recording medium that records such a program.
  • step S26 may be executed before step S25 if it is after step S24.
  • Step S26 may be executed in parallel with step S25 after step S24.
  • the image of the image file selected by the user's selection operation from the image files held in the background holding unit 107 is acquired as the background image (superposed image).
  • a composite image in which a visual image of the AR content is superimposed on the background image is displayed. Therefore, according to the present embodiment, the user performs an operation of selecting an image file in which a desired background image is stored from among the held image files, so that the AR content is superimposed on the desired background image.
  • An image (AR image) can be seen. As a result, the user can view the AR image of the desired background image anytime and anywhere.
  • the present inventors have solved the new idea of eliminating the limitation of the background image on which the AR content is superimposed and presenting the AR image in which the AR content is superimposed on the desired background image without limitation of place or time.
  • the present inventors are a new user who allows a user to select an image file in which a desired background image is stored from a plurality of image files when he / she wants to view an AR image.
  • the plurality of image files are image files in which a plurality of background images, each of which includes a marker, are stored so as to be a superimposed image.
  • Such a user interface has a completely different concept from the user interface of an existing AR application that is based on the assumption that AR content is superimposed on real-time video.
  • the following problems in the existing AR application can be solved.
  • a portable terminal with a camera is used as a device that captures and displays an AR image
  • the user must always shoot the real world on which AR content is superimposed on the portable terminal.
  • the user needs to always raise his arm. Therefore, the user cannot see the AR image with an easy posture.
  • the portable terminal is held by the user's hand, it is difficult to completely fix the position of the portable terminal. Therefore, the background image is blurred.
  • the image of the image file selected by the selection operation by the user is used as the background image (superimposed image), the user does not mind camera shake and has an easy posture. You can see the image.
  • the visual image of the AR content is set based on the marker detected from the background image, and the AR image placed at the position in the three-dimensional coordinate space is calculated based on the marker. Represents the field of view visually recognized from the background photographing position and the background photographing direction. Therefore, the user can recognize the situation as if the AR content existed in the real world shown in the desired background image by viewing the displayed AR image.
  • the image processing apparatus uses one content image among a plurality of content images obtained by photographing a real object from a plurality of directions as a visual image of the AR content, and uses an AR image (composite). Image).
  • FIG. 3 is a diagram conceptually illustrating a hardware configuration example of the image processing apparatus 10 according to the second embodiment.
  • the image processing apparatus 10 is a so-called computer.
  • the form of the image processing apparatus 10 such as a portable type or a stationary type is not limited.
  • the image processing apparatus 10 includes, for example, a CPU (Central Processing Unit) 11, a memory 12, an input / output interface (I / F) 13, a communication unit 14, and the like that are connected to each other via a bus.
  • the memory 12 is at least one of a RAM (Random Access Memory), a ROM (Read Only Memory), and a hard disk.
  • the communication unit 14 exchanges signals with other computers and devices.
  • a portable recording medium 18 or the like can be connected to the communication unit 14.
  • the input / output I / F 13 can be connected to the display device 15 and a user interface device such as the input device 16.
  • the display device 15 is generated based on drawing data processed by a CPU 11 or a GPU (Graphics Processing Unit) (not shown) such as an LCD (Liquid Crystal Display) or a CRT (Cathode Ray Tube) display. It is a device that displays a screen.
  • the display device 15 corresponds to the display unit 108 described above.
  • the input device 16 is a device that receives an input of a user operation, such as a keyboard and a mouse.
  • the display device 15 and the input device 16 may be integrated.
  • the display device 15 and the input device 16 may be realized as a touch panel, for example.
  • the display device 15 and the input device 16 may be connected to another computer that is communicably connected to the image processing device 10 via the communication unit 14.
  • the image processing apparatus 10 may have other hardware elements not shown in FIG.
  • the image processing apparatus 10 may have a camera corresponding to the imaging unit.
  • the image processing apparatus 10 can cause the background holding unit 31 to hold the image file of the background image captured by the camera. Then, the image processing apparatus 10 can capture a content image of a real object described later with the camera.
  • the hardware configuration of the image processing apparatus 10 is not limited.
  • FIG. 4 is a diagram conceptually illustrating a processing configuration example of the image processing apparatus 10 according to the second embodiment.
  • the image processing apparatus 10 according to the second embodiment includes an operation reception unit 21, a background acquisition unit 22, a marker detection unit 23, a reference calculation unit 24, a position determination unit 25, an image extraction unit 26, a size adjustment unit 27, and a synthesis processing unit. 28, a display control unit 29, and the like.
  • the image processing apparatus 10 further includes a background holding unit 31, a content holding unit 32, and the like.
  • These processing units are realized, for example, by executing a program stored in the memory 12 by the CPU 11.
  • the program is installed from a portable recording medium 18 such as a CD (Compact Disc) or a memory card or another computer on the network via the input / output I / F 13 or the communication unit 14, and the memory 12. May be stored.
  • a portable recording medium 18 such as a CD (Compact Disc) or a memory card or another computer on the network via the input / output I / F 13
  • the operation reception unit 21 corresponds to the operation reception unit 101 described above.
  • the operation reception unit 21 receives a user operation using the input device 16 for a screen displayed on the display device 15.
  • the operation accepting unit 21 further accepts a content image moving operation with respect to the composite image in addition to the image file selecting operation.
  • the selection operation includes, for example, an operation for selecting one image file from a list of a plurality of image files displayed on the display device 15 and a button operation for confirming the operation.
  • the movement operation is, for example, an operation of dragging and dropping or swiping the content image on the composite image displayed on the display device 15.
  • the content image is an image that is superimposed as a visual image of the AR content on the background image. The content image will be described in detail later.
  • the operation accepting unit 21 can accept an operation for selecting one real object to be displayed from among a plurality of real objects.
  • a real object is an object or person existing in the real world that is the source of AR content superimposed on a background image.
  • the operation accepting unit 21 can cause the display device 15 to display a list of a plurality of real objects related to content images held in the content holding unit 32 described later.
  • the operation reception part 21 can receive selection operation using the input device 16 with respect to the list
  • the background acquisition unit 22, the background holding unit 31, the marker detection unit 23, and the reference calculation unit 24 are the same as the background acquisition unit 102, the background holding unit 107, the marker detection unit 103, and the reference calculation unit 104 in the first embodiment described above. It is the same.
  • the position determination unit 25 determines the position in the three-dimensional coordinate space where the AR content to be displayed is virtually arranged, like the position determination unit 105 described above. Furthermore, when the movement operation is received by the operation reception unit 21, the position determination unit 25 determines a position after movement of the AR content on the three-dimensional coordinate space indicated by the movement operation. Specifically, the position determination unit 25 converts the moved position indicated by the moving operation on the composite image into a position in the three-dimensional coordinate space based on the marker actual information and image information. The position determination unit 25 determines the converted position as the position after movement of the AR content.
  • the content holding unit 32 holds a plurality of content images respectively associated with the content shooting directions obtained by shooting a real object corresponding to the AR content from a plurality of directions.
  • the content shooting direction is the direction in which the content image is shot.
  • the content holding unit 32 since the plurality of content images are used as visually recognized images of the AR content, it is desirable that the content holding unit 32 hold a large number of content images taken from all directions with a fine angular width. The reason is that the content image superimposed on the background image can be approximated to the visual image of the real object from the background shooting direction and the background shooting position of the background image.
  • FIG. 5 is a diagram conceptually showing an example of a plurality of content images held in the content holding unit 32 and associated with various content shooting directions.
  • the content shooting direction is shown in a two-axis coordinate space in the horizontal direction and the vertical direction.
  • the content holding unit 32 can also hold a plurality of content images each associated with a position coordinate indicating the content shooting direction.
  • each content image includes a marker and other backgrounds along with the real object.
  • the content holding unit 32 may hold a content image in which something other than a real object is captured.
  • the content holding unit 32 holds information on the size and shape of the marker in the real world when a marker appears in the content image to be held.
  • the content holding unit 32 holds information about the size of the real world of the real object when the marker is not captured in the content image to be held.
  • the content holding unit 32 can also hold a plurality of content images associated with various content shooting directions for each of the plurality of real objects.
  • the operation reception unit 21 can receive an operation of selecting one real object to be displayed from among a plurality of real objects.
  • the image extraction unit 26 holds the content image associated with the content shooting direction close to the direction from the position determined by the position determination unit 25 to the background shooting position calculated by the reference calculation unit 24 from the content holding unit 32. To extract. It is desirable that the content image associated with the content shooting direction that coincides with the direction from the position of the AR content in the three-dimensional coordinate space to the background shooting position is held in the content holding unit 32. The reason is that the content image matches the visual image of the real object from the shooting position of the background image. When such a matching content image is not held in the content holding unit 32, the image extraction unit 26 extracts a content image held in association with the closest content shooting direction.
  • the operation receiving unit 21 receives an operation of selecting one real object from the plurality of real objects.
  • the image extraction unit 26 extracts one content image from a plurality of content images associated with the selected real object that is retained in the content retention unit 32.
  • the position determination unit 25 determines a position after the movement.
  • the image extraction unit 26 is held in the content holding unit 32 in association with a content shooting direction that is close to the direction from the moved position determined by the position determination unit 25 to the background shooting position calculated by the reference calculation unit 24.
  • New content images are extracted.
  • the newly extracted content image is a visual image of a real object that exists at a position after the background image is moved from the shooting position.
  • the size adjustment unit 27 is extracted by the image extraction unit 26 based on the actual size of the real object, the real information and image information regarding the marker detected by the marker detection unit 23, and the position determined by the position determination unit 25.
  • the enlarged content image is enlarged or reduced.
  • the size adjusting unit 27 determines the enlargement ratio or reduction ratio of the content image so that the real object shown in the content image is superimposed on the real world shown in the background image in the size of the real world.
  • the size adjusting unit 27 can obtain the reduction ratio of the real world in the background image by using the real information and the image information regarding the marker.
  • the size adjustment unit 27 uses the real size of the real object, the image size in the content image, the reduction ratio of the real world in the background image, and the distance between the background shooting position and the AR content position. The enlargement or reduction ratio of the image can be determined.
  • the composition processing unit 28 is a composite image in which the visual image from the background photographing position and the background photographing direction of the AR content arranged at the position determined by the position determining unit 25 in the three-dimensional coordinate space is superimposed on the background image. (AR image) is generated. Specifically, the composition processing unit 28 superimposes the content image extracted by the image extraction unit 26 on the background image acquired by the background acquisition unit 22 as a visual image of the real object (AR content).
  • the composition processing unit 28 switches the content image to be superimposed on the background image to the content image newly extracted by the image extraction unit 26 in response to the movement operation received by the operation reception unit 21. Further, when the content image is enlarged or reduced by the size adjusting unit 27, the composition processing unit 28 superimposes the enlarged or reduced content image on the background image.
  • the display control unit 29 causes the display device 15 to display the composite image generated by the composition processing unit 28.
  • FIG. 6 is a flowchart illustrating an operation example of the image processing apparatus 10 according to the second embodiment.
  • the image processing apparatus 10 executes each processing step (that is, step).
  • a processing unit corresponding to the processing step to be executed may execute the processing step. 6
  • the same processing steps as those shown in FIG. 2 are denoted by the same reference numerals as those in FIG.
  • the image processing apparatus 10 accepts a user operation using the input device 16 (step S61).
  • the content of step S61 is the same as the processing content of the operation reception part 21.
  • FIG. When the selection operation is accepted (“selection” in step S62), the image processing apparatus 10 executes the processing steps from step S22 to step S26 as in FIG.
  • the image processing apparatus 10 extracts a content image from the content image held in the content holding unit 32 (step S63).
  • the image processing apparatus 10 may extract a content image associated with a content shooting direction that is close to the direction from the position determined in step S26 to the background shooting position calculated in step S25.
  • the content of step S63 is the same as the processing content of the image extraction unit 26.
  • step S64 the image processing apparatus 10 enlarges or reduces the content image extracted in step S63 so that the real object is superimposed on the real world in the size of the real world and reflected in the background image (step S64).
  • the content of step S64 is the same as the processing content of the size adjustment unit 27.
  • the image processing apparatus 10 generates a composite image by superimposing the content image enlarged or reduced in step S64 on the background image acquired in step S22 (step S65).
  • the content of step S65 is the same as the processing content of the composition processing unit 28.
  • the image processing device 10 causes the display device 15 to display the composite image generated in step S65 (step S66).
  • step S62 the image processing apparatus 10 determines the position after the movement of the AR content in the three-dimensional coordinate space indicated by the moving operation accepted in step S61. (Step S67).
  • the content of step S67 is the same as the processing content of the position determination unit 25.
  • the image processing apparatus 10 newly extracts a content image from the content image held in the content holding unit 32 (step S63).
  • the image processing apparatus 10 may extract a content image associated with a content shooting direction that is close to the direction from the moved position determined in step S67 to the background shooting position calculated in step S25. Thereafter, the image processing apparatus 10 executes step S64 and subsequent steps on the newly extracted content image.
  • the image processing apparatus 10 When content images of a plurality of real objects are held in the content holding unit 32, the image processing apparatus 10 performs an operation of selecting one real object to be displayed from the plurality of real objects in step S61. Can also be accepted. In this case, in step S63, the image processing apparatus 10 extracts one content image associated with the real object selected by the operation from the plurality of content images held in the content holding unit 32. .
  • the above operation in the case where content images of a plurality of real objects are held in the content holding unit 32 is not shown in FIG.
  • an image of an image file selected by a user's selection operation from among a plurality of image files is used as a background image (superimposed image).
  • a background image superimposed image
  • an AR image in which AR content is superimposed on a desired background image can be shown.
  • a plurality of content images obtained by photographing a real object from a plurality of directions and respectively associated with the content photographing directions are stored in advance. Then, an AR image (composite image) is generated using one content image among the plurality of held content images as a visual image of the AR content.
  • AR image composite image
  • the threshold for development and use of a platform using the AR technology can be reduced. Furthermore, the platform using the AR technology can be easily used by general users.
  • the held content image associated with the content shooting direction close to the direction from the AR content position to the background image shooting position in the three-dimensional coordinate space is the visual recognition of the real object.
  • the image is superimposed on the background image. Therefore, according to the second embodiment, the content image to be superimposed on the background image can be approximated to the visual image of the real object from the shooting position and shooting direction of the background image. Therefore, it is possible to enhance the realism of the space in which the real object is superimposed on the real world shown in the background image.
  • the position after movement in the three-dimensional coordinate space indicated by the movement operation is determined. Furthermore, a content image to be held that is associated with a content shooting direction that is close to the direction from the determined position after movement to the shooting position of the background image is newly extracted. The content image to be superimposed on the background image is switched to the newly extracted content image. Therefore, when a movement operation is performed on the content image on the AR image, the appearance of the real object on the AR image changes according to the position after the movement. As described above, according to the second embodiment, the user can move the real object as the AR content on the AR image while maintaining the reality.
  • the content image to be superimposed is enlarged or reduced so that the real object is superimposed on the real world in the background image with the size in the real world. Therefore, according to the second embodiment, the size relationship in the real world between the real object and the background reflected in the background image can be appropriately recognized by the user.
  • a plurality of content images associated with various content shooting directions can be held for each of the plurality of real objects.
  • the user can select not only a desired background image but also a desired real object as an AR content to be superimposed.
  • the image processing apparatus selectively operates a function of acquiring an image of an image file as a background image and a function of acquiring an image frame of a real-time video as a background image.
  • the image processing apparatus 10 according to the third embodiment will be described focusing on contents different from those of the second embodiment. In the following description, the same contents as those in the second embodiment are omitted as appropriate.
  • FIG. 7 is a diagram conceptually illustrating a hardware configuration example of the image processing apparatus 10 according to the third embodiment.
  • the hardware configuration of the image processing apparatus 10 according to the third embodiment may be the same as the hardware configuration of the image processing apparatus 10 according to the second embodiment.
  • the image processing apparatus 10 according to the third embodiment further includes a camera 17 that corresponds to an imaging unit in addition to the configuration of the second embodiment. That is, the camera 17 is further connected to the image processing apparatus 10 in the present embodiment.
  • the camera 17 is connected to the input / output I / F 13 or the communication unit 14.
  • the image processing apparatus 10 according to the third embodiment has the same functional configuration as that of the second embodiment shown in FIG.
  • the operation accepting unit 21 can also accept an operation for requesting acquisition of a real-time video from the camera 17 (hereinafter, a request operation). In response to accepting this operation, the operation accepting unit 21 can also activate the camera 17.
  • the background acquisition unit 22 sequentially acquires the real-time video image frame obtained from the camera 17 as a background image according to the received request operation. Thereby, the background image as the superimposed image is switched from the image of the image file to the image frame of the real-time video.
  • the display control unit 29 causes the display device 15 to display a real-time video in which the content image as described above is superimposed as a visual image.
  • FIG. 8 is a flowchart illustrating a part of an operation example of the image processing apparatus 10 according to the third embodiment.
  • the image processing apparatus 10 executes each processing process.
  • a processing unit corresponding to the processing step to be executed may execute the processing step.
  • FIG. 8 the same processing steps as those shown in FIG. 6 are denoted by the same reference numerals as in FIG.
  • the image processing apparatus 10 first executes step S61.
  • the image processing apparatus 10 receives an operation requesting acquisition of real-time video in step S61, the image processing apparatus 10 performs the operation shown in FIG.
  • the image processing apparatus 10 acquires an image frame of a real-time video obtained from the camera 17 as a background image (step S81). Thereafter, the processing steps after step S23 are executed on the background image acquired in step S81, as in the second embodiment described above.
  • the image processing apparatus 10 displays the real-time video on which the content image is superimposed as a visual image on the display device 15 by repeatedly executing the operation shown in FIG. 8 for each image frame of the real-time video obtained from the camera 17. be able to.
  • the image processing apparatus 10 does not have to use all the image frames constituting the real-time video as the background image.
  • the image processing apparatus 10 may perform thinning by removing some image frames from the image frames constituting the real-time video according to a predetermined method.
  • the image processing apparatus 10 may use the image frame after thinning as a background image. Further, the image processing apparatus 10 can omit the processing steps from step S23 to step S64 when the marker arrangement does not change between adjacent image frames.
  • the background image on which the AR content is superimposed is switched from the image of the image file to the image frame of the real-time video according to the operation by the user. Then, a held content image associated with a content shooting direction close to the direction from the AR content position to the shooting position of the image frame is superimposed on the image frame of the real-time image as a visual image of the real object. Is done.
  • video with which the content image was superimposed as a visual recognition image can be displayed on the display apparatus 15.
  • the first modification is a modification of the second embodiment of the present invention.
  • the third embodiment of the present invention can be modified as in this modification.
  • the user since the user can select a background image, the user may erroneously select an image file in which a background image without a marker is recorded.
  • the display control unit 29 causes the display device 15 to display that no marker is detected when no marker is detected from the background image by the marker detection unit 23.
  • the operation accepting unit 21 accepts an operation of reselecting an image file other than the image file in which no marker is detected.
  • the background acquisition unit 22 newly acquires an image of the image file newly selected by the reselection operation as a background image.
  • the user can recognize whether or not a background image that cannot superimpose the AR content is stored in the selected image file. Furthermore, the user can reselect the image file. Therefore, the user can perform an image file selection operation with peace of mind.
  • the second modification is a modification of the second embodiment of the present invention.
  • the third embodiment of the present invention can be modified as in this modification.
  • the time and place for generating the image file for storing the background image are arbitrary.
  • the user can hold a plurality of image files in which photographs showing a plurality of different places are recorded in the background holding unit 31 in order to use the plurality of image files as a background image. Under such circumstances, it may be more convenient to provide multiple types of markers.
  • the image processing apparatus 10 needs to know the actual information (shape information and size information) of the marker that appears in the background image.
  • FIG. 9 is a diagram conceptually illustrating a processing configuration example of the image processing apparatus 10A according to the second modification.
  • the image processing apparatus 10A according to the second modification further includes a marker information holding unit 35 in addition to the configurations of the second and third embodiments.
  • the marker information holding unit 35 is also realized in the same manner as other processing units.
  • another computer that is communicably connected to the image processing apparatus 10 ⁇ / b> A may have the marker information holding unit 35.
  • the marker information holding unit 35 holds marker information including shape information and size information for each of a plurality of marker types.
  • the marker detection unit 23 acquires shape information and size information of the specified type of marker from the marker information holding unit 35.
  • the marker detection unit 23 detects a marker from the background image acquired by the background acquisition unit 22 using the acquired shape information and size information.
  • the method for specifying the marker type is not limited.
  • the marker detection unit 23 can test whether or not a marker can be sequentially detected for each type of marker held in the marker information holding unit 35.
  • the marker detection unit 23 can cause the display device 15 to display a screen for allowing the user to select a marker type.
  • the marker detection unit 23 can also specify the marker type in response to an operation of selecting the marker type by the user.
  • the third modification is a modification of the third embodiment of the present invention.
  • the following processing unit of the image processing apparatus 10 according to the third embodiment may perform the following processing.
  • the operation reception unit 21 receives a snapshot operation in a state where a real-time video on which a visual image is superimposed is displayed on the display device 15.
  • the background acquisition unit 22 fixedly holds the image frame of the real-time video obtained by the camera 17 as the background image in response to reception of the snapshot operation by the operation reception unit 21. At this time, the background acquisition unit 22 can also stop the camera 17 after holding the image frame fixedly.
  • the display control unit 29 causes the display device 15 to display a composite image in which a content image is superimposed as a visual image on a fixedly held background image, instead of displaying a real-time video on which the visual image is superimposed.
  • the AR image on which the content image is superimposed can be displayed on the display device 15 by using the image frame obtained when the snapshot operation by the user is accepted as a fixed background image. .
  • the user can see the AR image in which the content image is superimposed on the image in which the real-time video is stopped at a desired timing.
  • the image processing apparatus 10 selectively operates a function of acquiring an image of an image file as a background image and a function of acquiring an image frame of a real-time video as a background image.
  • the image processing apparatus 10 may operate the function of acquiring the image frame of the real-time video as the background image without having the function of acquiring the image of the image file as the background image. In this case, the image processing apparatus 10 may not have the operation receiving unit 21.
  • FIG. 10A, FIG. 10B, and FIG. 10C are diagrams respectively showing examples of background images stored in the image file.
  • each background image includes a marker.
  • the same marker is included in each background image, but the marker included in each background image may not be the same as long as it is a marker known by the image processing apparatus 10.
  • the background holding unit 31 holds a plurality of image files for storing different background images.
  • the content holding unit 32 holds a content image.
  • Each content image held by the content holding unit 32 is associated with a unit vector representing the direction from the real object to the camera 17 when photographing the real object as information representing the content photographing direction.
  • the image processing apparatus 10 calculates a unit vector indicating a direction from the position of the AR content in the three-dimensional coordinate space to the shooting position of the background image.
  • the image processing apparatus 10 calculates the inner product of the unit vector calculated from the background image and each unit vector of each content image held in the content holding unit 32.
  • the image processing apparatus 10 specifies the unit vector of the content image that has the maximum calculated inner product value.
  • the image processing apparatus 10 extracts the held content image associated with the specified unit vector from the content holding unit 32.
  • the image processing apparatus 10 When synthesizing the content image and the background image, the image processing apparatus 10 (the synthesis processing unit 28) first generates a transparent plate-shaped three-dimensional object defined in the three-dimensional coordinate space. Also good. This three-dimensional object is a virtual object. The image of the three-dimensional object is superimposed on the background image.
  • the image processing apparatus 10 (the composition processing unit 28) may paste the content image extracted as described above as a texture on a plate-shaped three-dimensional object. As a result, the image processing apparatus 10 (the composition processing unit 28) superimposes the content image on the background image.
  • the image processing apparatus 10 determines the position of the three-dimensional object and the content image so that the image of the three-dimensional object to which the content image is pasted is substantially a visual image of the real object at the background shooting position. What is necessary is just to set the method to paste.
  • the method of superimposing the content image on the background image is not limited to the above example.
  • FIGS. 10A, 10B, and 10C are diagrams illustrating examples of a composite image (AR image) in which a content image is superimposed on the background image illustrated in FIGS. 10A, 10B, and 10C.
  • AR image a composite image
  • the superimposed content image is switched according to the shooting position and shooting direction of the background image.
  • the image of the real object in the superimposed content image has the same form as the real object installed in the space where the background image is captured, which is viewed from the shooting position and shooting direction of the background image.
  • FIGS. 12B and 12C are diagrams illustrating movement of AR content on an AR image.
  • the AR content is arranged on the marker.
  • FIG. 12B compared with the state shown in FIG. 12A, the AR content is moved backward with respect to the marker.
  • FIG. 12C compared with the state shown in FIG. 12A, the AR content is moved to the left with respect to the marker.
  • the user can view the AR image as shown in FIGS. 12B and 12C by performing an operation of moving the content image in a state where the AR image shown in FIG. 12A is displayed on the display device 15. .
  • FIG. 12B the content image is reduced in size compared to the content image shown in FIG. 12A according to the position of the AR content after movement.
  • FIG. 12C the content image is switched from the content image shown in FIG. 12A according to the position after movement.
  • An operation reception unit for receiving an image file selection operation;
  • a background acquisition unit for acquiring, as a background image, an image of an image file selected by a selection operation received by the operation reception unit, from a background holding unit for storing an image file storing a background image in which a marker is reflected;
  • a marker detection unit for detecting a marker from the background image acquired by the background acquisition unit;
  • a reference calculation unit that sets a three-dimensional coordinate space based on real information and image information regarding the detected marker, and calculates a background shooting position and a background shooting direction;
  • a position determining unit that determines a position in the three-dimensional coordinate space in which AR (Augmented Reality) content to be displayed is virtually arranged;
  • Display control for causing a display unit to display a composite image in which a visual image from the background photographing position and the background photographing direction of the AR content arranged at the determined position in the three-dimensional coordinate space is superimposed on the background image.
  • An image processing apparatus comprising: (Appendix 2) From a content holding unit that holds a plurality of content images obtained by shooting a real object corresponding to the AR content from a plurality of directions in association with each content shooting direction, from the position determined by the position determination unit, the background An image extraction unit for extracting a content image held in association with the content shooting direction close to the direction to the shooting position; A composition processing unit that superimposes the content image extracted by the image extraction unit on the background image as the visual image;
  • the operation accepting unit further accepts an operation for moving the content image with respect to the composite image,
  • the position determination unit determines a position after movement of the AR content in the three-dimensional coordinate space indicated by the accepted movement operation;
  • the image extraction unit newly extracts a content image held in the content holding unit in association with the content shooting direction close to the direction from the moved position determined by the position determination unit to the background shooting position.
  • the composition processing unit switches the content image to be superimposed on the background image to the newly extracted content image.
  • the image processing apparatus according to appendix 2. (Appendix 4) Enlarging or extracting the extracted content image based on the real size of the real object, the real information and the image information regarding the marker detected by the marker detection unit, and the position determined by the position determination unit Size adjustment part to reduce, Further comprising The composition processing unit superimposes the content image adjusted by the size adjusting unit on the background image as the visual image.
  • the image processing apparatus according to appendix 2 or 3.
  • the content holding unit holds the plurality of content images associated with each content shooting direction for each of a plurality of real objects
  • the operation accepting unit further accepts an operation of selecting one real object to be displayed from the plurality of real objects
  • the image extraction unit extracts one content image from the plurality of content images held in the content holding unit in relation to a real object selected by the accepted operation;
  • the image processing apparatus according to any one of appendices 2 to 4.
  • the display control unit causes the display unit to display that no marker is detected when the marker detection unit does not detect the marker from the background image
  • the operation accepting unit accepts an operation of reselecting an image file other than the image file in which the marker is not detected
  • the background acquisition unit newly acquires an image of an image file newly selected by the operation as the background image.
  • the image processing apparatus according to any one of appendices 1 to 5.
  • the operation accepting unit accepts an operation requesting acquisition of a real-time video from the imaging unit, and the background acquisition unit is the real-time video obtained from the imaging unit in response to the operation accepted by the operation accepting unit.
  • the display control unit causes the display unit to display a real-time video in which the content image is superimposed as the visual image.
  • the image processing apparatus according to any one of appendices 2 to 5. (Appendix 8)
  • the operation accepting unit accepts a snapshot operation in a state where a real-time video on which the visual image is superimposed is displayed on the display unit,
  • the background acquisition unit fixedly holds the image frame as the background image in response to reception of the snapshot operation by the operation reception unit,
  • the display control unit displays the composite image in which the content image is superimposed as the visual image on the fixedly held background image on the display unit, instead of displaying the real-time video on which the visual image is superimposed.
  • the marker detection unit acquires shape information and size information of one type of marker from a marker information holding unit that holds information of a plurality of types of markers each including shape information and size information, and the acquired shape information and Detecting the marker using size information;
  • the image processing apparatus according to any one of appendices 1 to 8.
  • (Appendix 11) The determined position in the three-dimensional coordinate space from a content holding unit that holds a plurality of content images obtained by shooting a real object corresponding to the AR content from a plurality of directions in association with each content shooting direction. A content image held in association with the content shooting direction that is close to the direction from the background shooting position to Superimposing the extracted content image on the background image as the visual image;
  • the image processing method further including: (Appendix 14) Receiving an operation of selecting one real object to be displayed from among a plurality of real objects; Further including The content holding unit holds the plurality of content images associated with the content shooting directions for each of the plurality of real objects, In the extraction of the content image, one content image is extracted from the plurality of content images held in the content holding unit in relation to the real object selected by the accepted operation.
  • the image processing method according to any one of appendices 11 to 13.
  • the image processing method according to any one of appendices 10 to 14, further including: (Appendix 16) Accepting an operation requesting acquisition of real-time video from the imaging unit, In response to the accepted operation, the image frame of the real-time video obtained from the imaging unit is sequentially acquired as the background image, Displaying the real-time video on which the content image is superimposed as the visual image on the display unit;
  • the image processing method according to any one of appendices 11 to 14, further including: (Appendix 17) Accepting a snapshot operation in a state where the real-time video on which the visual image is superimposed is displayed on the display unit, In response to accepting the snapshot operation, the image frame is fixedly held as the background image, Instead of displaying the real-time video on which the visual image is superimposed, the composite image in which the content image is superimposed as the visual image on the fixedly held background image is displayed on the display unit.
  • the image processing method according to appendix 16 further including: (Appendix 18) Obtaining shape information and size information of one type of marker from a marker information holding unit that holds information of a plurality of types of markers each including shape information and size information; Further including The detection of the marker detects the marker using the acquired shape information and size information. 18.
  • a background acquisition unit that sequentially acquires an image frame of a real-time video obtained from the imaging unit as a background image;
  • a marker detection unit for detecting a marker from the background image acquired by the background acquisition unit;
  • a reference calculation unit that sets a three-dimensional coordinate space based on real information and image information regarding the detected marker, and calculates a background shooting position and a background shooting direction;
  • a position determining unit that determines a position in the three-dimensional coordinate space in which AR (Augmented Reality) content to be displayed is virtually arranged; From a content holding unit that holds a plurality of content images obtained by shooting a real object corresponding to the AR content from a plurality of directions in association with each content shooting direction, from the position determined by the position determination unit, the background An image extraction unit for extracting a content image held in association with the content shooting direction close to the direction to the shooting position; A composition processing unit that superimposes the content image extracted by the image extraction unit on the background image as the visual image;
  • a display control unit that causes a
  • (Appendix 21) Obtain sequentially the image frame of the real-time video obtained from the imaging unit as a background image, Detecting a marker from the acquired background image; Based on real information and image information regarding the detected marker, a three-dimensional coordinate space is set, Based on the actual information and image information regarding the detected marker, calculate a background shooting position and a background shooting direction of the background image, Determining a position in the three-dimensional coordinate space where the AR (Augmented Reality) content to be displayed is virtually arranged; The determined position in the three-dimensional coordinate space from a content holding unit that holds a plurality of content images obtained by shooting a real object corresponding to the AR content from a plurality of directions in association with each content shooting direction.

Abstract

An objective of the present invention is to provide a technology whereby an AR image formed by superpositioning AR content on a desired background image is presented without restriction on place or time. Provided is an image processing device comprising: a manipulation acceptance means for accepting a selection manipulation on an image file; a background acquisition means for acquiring, from a background retaining means for retaining image files, an image of the image file selected with the selection manipulation as a background image; a marker detection means for detecting a marker from the acquired background image; a reference computation means for setting a three-dimensional coordinate space and computing a background photography location and a background photography direction on the basis of reality information and image information relating to the detected marker; a location determination means for determining a location in the three-dimensional coordinate space where AR content to be displayed is virtually positioned; and a display control means for causing a display means to display a composite image formed by superpositioning on the background image a viewing image, to be seen from the background photography location and background photography direction, of the AR content positioned at the determined location in the three-dimensional coordinate space.

Description

画像処理装置、画像処理方法及びプログラム記憶媒体Image processing apparatus, image processing method, and program storage medium
 本発明は、拡張現実(Augmented Reality;AR)の画像処理技術に関する。 The present invention relates to augmented reality (AR) image processing technology.
 現在、拡張現実と呼ばれる技術を用いた様々なアプリケーションが提供されている。AR技術は、スマートフォン等の携帯機器のカメラやヘッドマウントディスプレイ(HMD)を介して得られる現実世界に3次元コンテンツや2次元コンテンツ(文字や図柄など)を重畳的に表示させる。例えば、下記非特許文献1には、現実世界に重畳させる3次元コンテンツを現実世界に存在する現実オブジェクトを実写した画像で生成する手法が提案されている。また、当該3次元コンテンツには、実写画像ではなく、コンピュータグラフィック(CG)技術を用いて生成される仮想3次元モデルが利用される形態もある。 Currently, various applications using technology called augmented reality are being provided. The AR technology displays three-dimensional content and two-dimensional content (such as characters and designs) in a superimposed manner in the real world obtained via a camera or a head-mounted display (HMD) of a mobile device such as a smartphone. For example, the following Non-Patent Document 1 proposes a method of generating a three-dimensional content to be superimposed on the real world with an image obtained by actually capturing a real object existing in the real world. In addition, there is a form in which the virtual three-dimensional model generated by using computer graphic (CG) technology is used for the three-dimensional content, instead of the photographed image.
 以降、現実世界に3次元コンテンツや2次元データが重畳された合成画像はAR画像と表記される場合もある。また、3次元コンテンツや2次元コンテンツが重畳される現実世界を写す画像が背景画像と表記される場合もある。また、この背景画像に重畳される3次元コンテンツや2次元コンテンツがARコンテンツと表記される場合もある。 Hereinafter, a composite image in which 3D content and 2D data are superimposed on the real world may be referred to as an AR image. In addition, an image of the real world on which 3D content or 2D content is superimposed may be described as a background image. In addition, 3D content and 2D content superimposed on the background image may be referred to as AR content.
 下記非特許文献2は、AR技術を用いた、移動ロボットの遠隔操縦システムを提案する。この提案手法では、ロボットの3次元モデルがAR技術を用いて拡張自由視点画像に重畳された画像が、HMDを介して操縦者に提示される。この拡張自由視点画像は、移動ロボットにより複数地点で取得された全方位画像群および距離画像群、並びに、当該取得時のロボットの位置及び姿勢に基づいて、生成される。また、下記特許文献1には、重畳する3次元コンテンツの生成手法として、注目物体の3次元表示をするための変換画像を、加速度センサ等の専用の機材を用いることなく、単一のカメラで生成する手法が提案されている。この提案手法では、注目物体を含む複数の画像についてカメラの回転軸を構成する複数の画像座標をユーザに指定させることによって、指定された画像座標が一致するように、変換対象画像が幾何変換される。 The following Non-Patent Document 2 proposes a remote control system for a mobile robot using AR technology. In this proposed method, an image in which a three-dimensional model of a robot is superimposed on an extended free viewpoint image using AR technology is presented to the operator via the HMD. This extended free viewpoint image is generated based on the omnidirectional image group and distance image group acquired at a plurality of points by the mobile robot, and the position and orientation of the robot at the time of acquisition. Further, in Patent Document 1 below, as a method for generating three-dimensional content to be superimposed, a converted image for three-dimensional display of a target object can be obtained with a single camera without using a dedicated device such as an acceleration sensor. A generation method has been proposed. In this proposed method, the conversion target image is geometrically transformed so that the specified image coordinates match by allowing the user to specify a plurality of image coordinates constituting the rotation axis of the camera for a plurality of images including the target object. The
国際公開第2011/148595号International Publication No. 2011/148595
 上述の提案手法や既存のアプリケーションでは、3次元コンテンツや2次元コンテンツが重畳される背景画像には、リアルタイムに撮像装置から得られるリアルタイム映像が利用される。これにより、ユーザは、撮像装置がリアルタイムに写す現実世界にあたかもARコンテンツが存在するようなAR画像を視認することができる。 In the above-described proposed method and existing applications, a real-time video obtained from an imaging device in real time is used as a background image on which 3D content or 2D content is superimposed. As a result, the user can visually recognize an AR image in which AR content exists in the real world captured by the imaging apparatus in real time.
 しかし、上述のような手法では、ARコンテンツを重畳させたい現実世界が、そのときの撮像装置の存在場所に制限されてしまう。例えば、ユーザが、所持するカメラ付き携帯端末で、自身の部屋にARコンテンツが重畳されたAR画像を見ることができるのは、ユーザがその携帯端末を持って自身の部屋にいるときだけである。ユーザは、電車の中等のようなユーザの部屋以外の場所では、そのAR画像を見ることはできない。 However, with the method as described above, the real world where the AR content is to be superimposed is limited to the location of the imaging device at that time. For example, a user can view an AR image in which AR content is superimposed on his / her room with his / her camera-equipped mobile terminal only when the user is in his / her room with the mobile terminal. . The user cannot see the AR image in a place other than the user's room such as in a train.
 本発明は、このような事情に鑑みてなされたものである。本発明の目的の一つは、場所及び時間の制限なく、所望の背景画像に、ARコンテンツが重畳されたAR画像を提示する技術を提供することにある。 The present invention has been made in view of such circumstances. One of the objects of the present invention is to provide a technique for presenting an AR image in which AR content is superimposed on a desired background image without restrictions on place and time.
 本発明の各側面では、上述した課題を解決するために、それぞれ以下の構成を採用する。 In each aspect of the present invention, the following configurations are adopted in order to solve the above-described problems.
 本発明の第1の側面は、画像処理装置に関する。本発明の第1の側面に係る画像処理装置は、画像ファイルの選択操作を受け付ける操作受付手段と、マーカが写る背景画像を保存する画像ファイルを保持する背景保持手段から、操作受付手段により受け付けられる選択操作によって選択された画像ファイルの画像を背景画像として取得する背景取得手段と、背景取得手段によって取得される背景画像からマーカを検出するマーカ検出手段と、検出されたマーカに関する現実情報及び画像情報に基づいて、3次元座標空間を設定し、かつ、背景撮影位置及び背景撮影方向を算出する基準算出手段と、表示対象のARコンテンツを仮想的に配置する3次元座標空間上の位置を決定する位置決定手段と、3次元座標空間の決定された位置に配置されるARコンテンツの背景撮影位置及び背景撮影方向からの視認画像が背景画像に重畳された合成画像を表示手段に表示させる表示制御手段と、を有する。 The first aspect of the present invention relates to an image processing apparatus. An image processing apparatus according to a first aspect of the present invention is received by an operation receiving unit from an operation receiving unit that receives an image file selection operation and a background holding unit that stores an image file that stores a background image in which a marker is captured. Background acquisition means for acquiring an image of the image file selected by the selection operation as a background image, marker detection means for detecting a marker from the background image acquired by the background acquisition means, and actual information and image information regarding the detected marker And a reference calculation means for calculating a background shooting position and a background shooting direction, and a position in the three-dimensional coordinate space where the AR content to be displayed is virtually arranged is determined. Position determining means, background shooting position and background shooting method of AR content arranged at determined position in three-dimensional coordinate space Viewed image from having a display control means for displaying on the display means a composite image superimposed on the background image.
 本発明の第2の側面は、画像処理方法に関する。本発明の第2の側面に係る画像処理方法は、画像ファイルの選択操作を受け付け、マーカが写る背景画像を保存する画像ファイルを保持する背景保持手段から、当該受け付けられた選択操作によって選択された画像ファイルの画像を背景画像として取得し、取得された背景画像からマーカを検出し、検出されたマーカに関する現実情報及び画像情報に基づいて、3次元座標空間を設定し、検出されたマーカに関する現実情報及び画像情報に基づいて、背景撮影位置及び背景撮影方向を算出し、表示対象のARコンテンツを仮想的に配置する3次元座標空間上の位置を決定し、3次元座標空間の決定された位置に配置されるARコンテンツの背景撮影位置及び背景撮影方向からの視認画像が背景画像に重畳された合成画像を表示手段に表示させる、ことを含む。 The second aspect of the present invention relates to an image processing method. The image processing method according to the second aspect of the present invention receives an image file selection operation, and is selected by the received selection operation from a background holding unit that holds an image file that stores a background image in which a marker is captured. An image of an image file is acquired as a background image, a marker is detected from the acquired background image, a three-dimensional coordinate space is set based on the actual information and image information regarding the detected marker, and the reality regarding the detected marker Based on the information and the image information, the background photographing position and the background photographing direction are calculated, the position on the three-dimensional coordinate space where the AR content to be displayed is virtually arranged is determined, and the determined position in the three-dimensional coordinate space Display on the display means a composite image in which the visual image from the background shooting position and the background shooting direction of the AR content arranged on the background image is superimposed on the background image. , Including that.
 本発明の第3の側面は、プログラムに関する。本発明の第3の側面に係るプログラムは、コンピュータを、画像ファイルの選択操作を受け付ける操作受付手段と、マーカが写る背景画像を保存する画像ファイルを保持する背景保持部から、前記操作受付手段により受け付けられる選択操作で選択された画像ファイルの画像を背景画像として取得する背景取得手段と、前記背景取得手段によって取得される前記背景画像からマーカを検出するマーカ検出手段と、前記検出されたマーカに関する現実情報及び画像情報に基づいて、3次元座標空間を設定し、かつ、背景撮影位置及び背景撮影方向を算出する基準算出手段と、表示対象のAR(Augmented Reality)コンテンツを仮想的に配置する前記3次元座標空間上の位置を決定する位置決定手段と、前記3次元座標空間の前記決定された位置に配置される前記ARコンテンツの前記背景撮影位置及び前記背景撮影方向からの視認画像が前記背景画像に重畳された合成画像を表示手段に表示させる表示制御手段と、して動作させる。
 本発明の第3側面は、上記プログラムを記録したコンピュータが読み取り可能な記録媒体によっても実現される。この記録媒体は、非一時的な有形の媒体を含む。
The third aspect of the present invention relates to a program. According to a third aspect of the present invention, there is provided a program comprising: an operation receiving unit that receives a selection operation of an image file; and a background holding unit that holds an image file that stores a background image in which a marker is captured. A background acquisition unit that acquires an image of an image file selected by an accepted selection operation as a background image, a marker detection unit that detects a marker from the background image acquired by the background acquisition unit, and the detected marker The reference calculation means for setting the three-dimensional coordinate space based on the real information and the image information and calculating the background photographing position and the background photographing direction and the AR (Augmented Reality) content to be displayed are virtually arranged. A position determining means for determining a position in the three-dimensional coordinate space; and a position determining unit arranged at the determined position in the three-dimensional coordinate space. Display control means for viewing images displayed on the display means a composite image superimposed on the background image from the background photographing position and the background photographing direction of the AR content is to operate with.
The third aspect of the present invention is also realized by a computer-readable recording medium that records the program. This recording medium includes a non-transitory tangible medium.
 上記各側面によれば、場所や時間の制限なく、所望の背景画像にARコンテンツが重畳されたAR画像を提示する技術を提供することができる。 According to each aspect described above, it is possible to provide a technique for presenting an AR image in which AR content is superimposed on a desired background image without any restrictions on place or time.
図1は、本発明の第1の実施形態における画像処理装置の処理構成例を概念的に示す図である。FIG. 1 is a diagram conceptually illustrating a processing configuration example of an image processing apparatus according to the first embodiment of the present invention. 図2は、本発明の第1の実施形態における画像処理装置の動作例を示すフローチャートである。FIG. 2 is a flowchart showing an operation example of the image processing apparatus according to the first embodiment of the present invention. 図3は、本発明の第2の実施形態における画像処理装置のハードウェア構成例を概念的に示す図である。FIG. 3 is a diagram conceptually illustrating a hardware configuration example of an image processing apparatus according to the second embodiment of the present invention. 図4は、本発明の第2の実施形態における画像処理装置の処理構成例を概念的に示す図である。FIG. 4 is a diagram conceptually illustrating a processing configuration example of the image processing apparatus according to the second embodiment of the present invention. 図5は、コンテンツ保持部に保持される、さまざまなコンテンツ撮影方向に関連付けられた複数のコンテンツ画像の例を概念的に示す図である。FIG. 5 is a diagram conceptually illustrating an example of a plurality of content images held in the content holding unit and associated with various content shooting directions. 図6は、本発明の第2の実施形態における画像処理装置の動作例を示すフローチャートである。FIG. 6 is a flowchart showing an operation example of the image processing apparatus according to the second embodiment of the present invention. 図7は、本発明の第3の実施形態における画像処理装置のハードウェア構成例を概念的に示す図である。FIG. 7 is a diagram conceptually illustrating a hardware configuration example of an image processing apparatus according to the third embodiment of the present invention. 図8は、本発明の第3の実施形態における画像処理装置の動作例の一部を示すフローチャートである。FIG. 8 is a flowchart showing a part of an operation example of the image processing apparatus according to the third embodiment of the present invention. 図9は、本発明の第3の変形例における画像処理装置の処理構成例を概念的に示す図である。FIG. 9 is a diagram conceptually illustrating a processing configuration example of the image processing apparatus according to the third modification of the present invention. 図10Aは、画像ファイルに保存される背景画像の例を示す図である。FIG. 10A is a diagram illustrating an example of a background image stored in an image file. 図10Bは、画像ファイルに保存される背景画像の例を示す図である。FIG. 10B is a diagram illustrating an example of a background image stored in an image file. 図10Cは、画像ファイルに保存される背景画像の例を示す図である。FIG. 10C is a diagram illustrating an example of a background image stored in an image file. 図11Aは、図10Aに示される背景画像にコンテンツ画像が重畳された合成画像(AR画像)の例を示す図である。FIG. 11A is a diagram showing an example of a composite image (AR image) in which a content image is superimposed on the background image shown in FIG. 10A. 図11Bは、図10Bに示される背景画像にコンテンツ画像が重畳された合成画像(AR画像)の例を示す図である。FIG. 11B is a diagram showing an example of a composite image (AR image) in which a content image is superimposed on the background image shown in FIG. 10B. 図11Cは、図10Cに示される背景画像にコンテンツ画像が重畳された合成画像(AR画像)の例を示す図である。FIG. 11C is a diagram illustrating an example of a composite image (AR image) in which a content image is superimposed on the background image illustrated in FIG. 10C. 図12Aは、AR画像上でのARコンテンツの移動を示す図である。FIG. 12A is a diagram illustrating movement of AR content on an AR image. 図12Bは、AR画像上でのARコンテンツの移動を示す図である。FIG. 12B is a diagram illustrating movement of AR content on an AR image. 図12Cは、AR画像上でのARコンテンツの移動を示す図である。FIG. 12C is a diagram illustrating movement of AR content on an AR image.
 以下、本発明の実施の形態について説明する。なお、以下に挙げる実施形態は例示であり、本発明は以下の実施形態の構成に限定されない。 Hereinafter, embodiments of the present invention will be described. In addition, embodiment mentioned below is an illustration and this invention is not limited to the structure of the following embodiment.
 [第1の実施形態]
 図1は、本発明の第1の実施形態における画像処理装置の処理構成例を概念的に示す図である。図1に示されるように、画像処理装置100は、操作受付部101、背景取得部102、マーカ検出部103、基準算出部104、位置決定部105、及び表示制御部106等を有する。画像処理装置100は、後述する詳細な実施形態における画像処理装置10と同様のハードウェア構成を有する。画像処理装置100は、その画像処理装置10と同様にプログラムによって制御されることにより、上述の各処理部が実現される。また、画像処理装置100は、背景保持部107及び表示部108を有してもよい。但し、画像処理装置100と通信可能に接続される他のコンピュータ(図示せず)が、背景保持部107及び表示部108を有していてもよい。表示部108は、LCD(Liquid Crystal Display)やCRT(Cathode Ray Tube)ディスプレイ等のような、例えば描画データに基づいて生成された画面を表示する装置である。
[First Embodiment]
FIG. 1 is a diagram conceptually illustrating a processing configuration example of an image processing apparatus according to the first embodiment of the present invention. As illustrated in FIG. 1, the image processing apparatus 100 includes an operation reception unit 101, a background acquisition unit 102, a marker detection unit 103, a reference calculation unit 104, a position determination unit 105, a display control unit 106, and the like. The image processing apparatus 100 has the same hardware configuration as the image processing apparatus 10 in a detailed embodiment described later. Similar to the image processing apparatus 10, the image processing apparatus 100 is controlled by a program, thereby realizing the above-described processing units. In addition, the image processing apparatus 100 may include a background holding unit 107 and a display unit 108. However, another computer (not shown) connected to be communicable with the image processing apparatus 100 may include the background holding unit 107 and the display unit 108. The display unit 108 is a device that displays a screen generated based on drawing data, such as an LCD (Liquid Crystal Display) or a CRT (Cathode Ray Tube) display.
 操作受付部101は、画像ファイルの選択操作を受け付ける。 The operation reception unit 101 receives an image file selection operation.
 背景取得部102は、マーカが写る背景画像を保存する画像ファイルを保持する背景保持部107から、操作受付部101により受け付けられる選択操作によって選択された画像ファイルの画像を背景画像として取得する。ここで、画像ファイルには、動画データ又は静止画データが保存される。背景取得部102は、動画データ又は静止画データから静止画情報としての背景画像を取得する。 The background acquisition unit 102 acquires, as a background image, the image of the image file selected by the selection operation received by the operation reception unit 101 from the background holding unit 107 that stores the image file that stores the background image in which the marker is reflected. Here, moving image data or still image data is stored in the image file. The background acquisition unit 102 acquires a background image as still image information from moving image data or still image data.
 マーカ検出部103は、背景取得部102で取得される背景画像から、マーカを検出する。ここで、マーカは、現実世界に配置されている物や、現実世界に配置されている物に描かれている画像である。マーカは、ARマーカ等と呼ばれる。但し、本実施形態では、このマーカから、参照方向によらず、或る基準点及びその基準点からの相互に直交する3つの方向を一定に得ることができるのであれば、このマーカの具体的形態は制限されない。例えば、マーカ検出部103は、マーカの現実情報(形状情報等)を予め保持している。そして、マーカ検出部103は、この現実情報に基づいて、背景画像からマーカを検出する。 The marker detection unit 103 detects a marker from the background image acquired by the background acquisition unit 102. Here, the marker is an image drawn on an object arranged in the real world or an object arranged in the real world. The marker is called an AR marker or the like. However, in the present embodiment, if a certain reference point and three directions perpendicular to each other from the reference point can be obtained from this marker regardless of the reference direction, the specific of this marker can be obtained. The form is not limited. For example, the marker detection unit 103 holds in advance marker actual information (such as shape information). And the marker detection part 103 detects a marker from a background image based on this real information.
 基準算出部104は、検出されたマーカに関する現実情報及び画像情報に基づいて、3次元座標空間を設定し、かつ、背景撮影位置及び背景撮影方向を算出する。マーカに関する現実情報は、マーカの現実世界における既知の情報を意味する。マーカに関する現実情報は、例えば、マーカの実際の形状を示す。また、マーカに関する画像情報は、マーカの背景画像内における情報を意味する。マーカに関する画像情報は、マーカの背景画像内の形状を示す。基準算出部104は、検出されたマーカから認識される基準点に応じた位置に3次元座標空間を設定する。3次元座標空間が設定される位置は、特に制限されない。また、基準算出部104は、マーカに関する現実情報と画像情報との比較により、背景撮影位置及び背景撮影方向を算出する。背景撮影位置及び背景撮影方向は、その画像ファイルの背景画像を撮影した撮像装置の位置及び向き(姿勢)を意味する。 The reference calculation unit 104 sets a three-dimensional coordinate space based on the detected real information and image information about the marker, and calculates the background shooting position and the background shooting direction. The actual information related to the marker means known information in the real world of the marker. The actual information regarding the marker indicates, for example, the actual shape of the marker. Moreover, the image information regarding a marker means the information in the background image of a marker. The image information regarding the marker indicates the shape of the marker in the background image. The reference calculation unit 104 sets a three-dimensional coordinate space at a position corresponding to the reference point recognized from the detected marker. The position where the three-dimensional coordinate space is set is not particularly limited. In addition, the reference calculation unit 104 calculates the background shooting position and the background shooting direction by comparing the actual information about the marker and the image information. The background shooting position and the background shooting direction mean the position and orientation (posture) of the imaging apparatus that shot the background image of the image file.
 位置決定部105は、表示対象のARコンテンツを仮想的に配置する、上記3次元座標空間上の位置を決定する。位置決定部105は、この位置を任意に又は固定的に決定することができる。表示対象のARコンテンツは、上記非特許文献1及び2で提案される3次元モデルであってもよい。表示対象のARコンテンツは、現実オブジェクトを複数の方向から撮影して得られた複数のコンテンツ画像を切り替え表示することで模擬される3次元コンテンツであってもよい。表示対象のARコンテンツは、平面的な2次元コンテンツであってもよい。 The position determination unit 105 determines the position in the three-dimensional coordinate space where the AR content to be displayed is virtually arranged. The position determination unit 105 can determine this position arbitrarily or fixedly. The AR content to be displayed may be a three-dimensional model proposed in Non-Patent Documents 1 and 2 above. The AR content to be displayed may be three-dimensional content that is simulated by switching and displaying a plurality of content images obtained by photographing a real object from a plurality of directions. The AR content to be displayed may be planar two-dimensional content.
 表示制御部106は、3次元座標空間における上記決定された位置に配置されるARコンテンツの、背景撮影位置及び背景撮影方向からの視認画像が、上記背景画像に重畳された合成画像を、表示部108に表示させる。ARコンテンツが上述の3次元モデル又は2次元コンテンツである場合には、表示制御部106は、3次元モデル又は2次元コンテンツの持つ情報に基づいて、周知の手法を用いて当該視認画像を得ることができる。ARコンテンツが、現実オブジェクトを複数の方向から撮影して得られた複数のコンテンツ画像を切り替え表示することで模擬される3次元コンテンツである場合の、当該視認画像の取得手法は、詳細な実施形態として後述する。 The display control unit 106 displays a composite image in which a visual image from the background shooting position and the background shooting direction of the AR content arranged at the determined position in the three-dimensional coordinate space is superimposed on the background image. 108 is displayed. When the AR content is the above-described three-dimensional model or two-dimensional content, the display control unit 106 obtains the visual image using a known method based on the information of the three-dimensional model or the two-dimensional content. Can do. A method for acquiring the visual image when the AR content is a three-dimensional content that is simulated by switching and displaying a plurality of content images obtained by photographing a real object from a plurality of directions is described in detail. Will be described later.
 図2は、本実施形態における画像処理装置100の動作例を示すフローチャートである。本実施形態における画像処理方法は、図2に示されるように、画像処理装置100のような少なくとも1つのコンピュータにより実行される。そのコンピュータは、後述される画像処理装置10と同様のハードウェア構成を備えていればよい。その画像処理方法は、ステップS21からステップS27までの処理工程を含む。 FIG. 2 is a flowchart showing an operation example of the image processing apparatus 100 in the present embodiment. The image processing method in the present embodiment is executed by at least one computer such as an image processing apparatus 100 as shown in FIG. The computer may have a hardware configuration similar to that of the image processing apparatus 10 described later. The image processing method includes processing steps from step S21 to step S27.
 ステップS21では、当該コンピュータは、画像ファイルの選択操作を受け付ける。 In step S21, the computer accepts an image file selection operation.
 ステップS22では、当該コンピュータは、マーカがそれぞれ背景画像を保存する画像ファイルを保持する背景保持部107から、ステップS21で受け付けられた選択操作で選択された画像ファイルの画像を背景画像として取得する。マーカ及び画像ファイルについては上述したとおりである。ステップS22の内容は、背景取得部102の処理内容と同様である。 In step S22, the computer acquires the image of the image file selected by the selection operation accepted in step S21 as the background image from the background holding unit 107 that holds the image files in which the markers each store the background image. The marker and the image file are as described above. The content of step S22 is the same as the processing content of the background acquisition unit 102.
 ステップS23では、当該コンピュータは、ステップS22で取得された背景画像からマーカを検出する。ステップS23の内容は、マーカ検出部103の処理内容と同様である。 In step S23, the computer detects a marker from the background image acquired in step S22. The content of step S23 is the same as the processing content of the marker detection unit 103.
 ステップS24では、当該コンピュータは、ステップS23で検出されたマーカに関する現実情報及び画像情報に基づいて、3次元座標空間を設定する。
 ステップS25では、当該コンピュータは、ステップS23で検出されたマーカに関する現実情報及び画像情報に基づいて、背景撮影位置及び背景撮影方向を算出する。ステップS24及びステップS25の内容は、基準算出部104の処理内容と同様である。
In step S24, the computer sets a three-dimensional coordinate space based on the real information and image information regarding the marker detected in step S23.
In step S25, the computer calculates a background shooting position and a background shooting direction based on the actual information and the image information regarding the marker detected in step S23. The contents of steps S24 and S25 are the same as the contents of processing of the reference calculation unit 104.
 ステップS26では、当該コンピュータは、表示対象のARコンテンツを仮想的に配置する上記3次元座標空間上の位置を決定する。ステップS26の内容は、位置決定部105の処理内容と同様である。 In step S26, the computer determines a position on the three-dimensional coordinate space where the AR content to be displayed is virtually arranged. The content of step S26 is the same as the processing content of the position determination unit 105.
 ステップS27では、当該コンピュータは、配置されるARコンテンツの視認画像が、背景画像に重畳された合成画像を、表示部108に表示させる。ARコンテンツが配置される位置は、ステップS24で設定された3次元座標空間における、ステップS26で決定された位置である。ARコンテンツの視認画像は、ステップS25で算出された背景撮影位置及び背景撮影方向からの、前述の位置に配置されるARコンテンツの視認画像である。背景画像は、ステップS22で取得された背景画像である。ステップS27の内容は、表示制御部106の処理内容と同様である。 In step S27, the computer causes the display unit 108 to display a composite image in which the visual image of the AR content to be arranged is superimposed on the background image. The position where the AR content is arranged is the position determined in step S26 in the three-dimensional coordinate space set in step S24. The visual image of the AR content is a visual image of the AR content arranged at the above-described position from the background shooting position and the background shooting direction calculated in step S25. The background image is the background image acquired in step S22. The content of step S27 is the same as the processing content of the display control unit 106.
 画像処理方法の実行主体である少なくとも1つのコンピュータが、上記表示部108及び上記背景保持部107を有していてもよい。この少なくとも1つのコンピュータと通信可能に接続される他のコンピュータ(図示せず)が、上記表示部108及び上記背景保持部107を有していてもよい。また、本発明の各実施形態は、上述の画像処理方法を少なくとも1つのコンピュータに実行させるプログラムによって実現されていてもよい。本発明の各実施形態は、このようなプログラムを記録した当該コンピュータが読み取り可能な記録媒体によって実現されていてもよい。 At least one computer that is the execution subject of the image processing method may include the display unit 108 and the background holding unit 107. Another computer (not shown) connected to be communicable with the at least one computer may include the display unit 108 and the background holding unit 107. In addition, each embodiment of the present invention may be realized by a program that causes at least one computer to execute the above-described image processing method. Each embodiment of the present invention may be realized by a computer-readable recording medium that records such a program.
 また、本実施形態における画像処理方法の処理工程の実行順は、図2に示される順番に制限されない。例えば、ステップS26は、ステップS24より後であれば、ステップS25の前に実行されてもよい。ステップS26は、ステップS24より後に、ステップS25と並列に実行されてもよい。 Further, the execution order of the processing steps of the image processing method in the present embodiment is not limited to the order shown in FIG. For example, step S26 may be executed before step S25 if it is after step S24. Step S26 may be executed in parallel with step S25 after step S24.
 このように、本実施形態では、背景保持部107に保持される画像ファイルの中からユーザによる選択操作で選択された画像ファイルの画像が背景画像(被重畳画像)として取得される。そして、この背景画像に対してARコンテンツの視認画像が重畳された合成画像が表示される。従って、本実施形態によれば、ユーザは、保持される画像ファイルの中から所望の背景画像を保存する画像ファイルを選択する操作を行うことによって、所望の背景画像にARコンテンツが重畳された合成画像(AR画像)を見ることができる。その結果、ユーザは、いつでもどこでも、所望の背景画像のAR画像を見ることができる。 Thus, in this embodiment, the image of the image file selected by the user's selection operation from the image files held in the background holding unit 107 is acquired as the background image (superposed image). A composite image in which a visual image of the AR content is superimposed on the background image is displayed. Therefore, according to the present embodiment, the user performs an operation of selecting an image file in which a desired background image is stored from among the held image files, so that the AR content is superimposed on the desired background image. An image (AR image) can be seen. As a result, the user can view the AR image of the desired background image anytime and anywhere.
 このように、本発明者らは、ARコンテンツを重畳させる背景画像の制限を解消し、場所や時間の制限なく、所望の背景画像にARコンテンツが重畳されたAR画像を提示するという新たな発想を見出した。そして、本発明者らは、上述のように、AR画像を見たいときに、ユーザに、複数の画像ファイルの中から所望の背景画像が保存されている画像ファイルを選択させるという、新たなユーザインタフェースを見出した。それらの複数の画像ファイルは、被重畳画像となり得るように、マーカがそれぞれ写る、複数の背景画像が保存される画像ファイルである。このようなユーザインタフェースは、リアルタイム映像にARコンテンツを重畳させることを前提とする既存のARアプリケーションのユーザインタフェースとはその発想が全く異なる。 As described above, the present inventors have solved the new idea of eliminating the limitation of the background image on which the AR content is superimposed and presenting the AR image in which the AR content is superimposed on the desired background image without limitation of place or time. I found. Then, as described above, the present inventors are a new user who allows a user to select an image file in which a desired background image is stored from a plurality of image files when he / she wants to view an AR image. Found the interface. The plurality of image files are image files in which a plurality of background images, each of which includes a marker, are stored so as to be a superimposed image. Such a user interface has a completely different concept from the user interface of an existing AR application that is based on the assumption that AR content is superimposed on real-time video.
 本実施形態におけるこのようなユーザインタフェースによれば、既存のARアプリケーションにおける次のような問題点も解消することができる。AR画像を撮影及び表示する装置としてカメラ付きの携帯端末が用いられる場合、ユーザは、その携帯端末でARコンテンツが重畳される現実世界を常に撮影しなければならない。これでは、AR画像を見るために、ユーザは、常に腕を上げている必要がある。そのため、ユーザは、楽な姿勢でAR画像を見ることができない。更に、携帯端末はユーザの手によって保持されるので、携帯端末の位置を完全に固定することは困難である。そのため、背景画像がぶれてしまう。 According to such a user interface in the present embodiment, the following problems in the existing AR application can be solved. When a portable terminal with a camera is used as a device that captures and displays an AR image, the user must always shoot the real world on which AR content is superimposed on the portable terminal. In this case, in order to view the AR image, the user needs to always raise his arm. Therefore, the user cannot see the AR image with an easy posture. Furthermore, since the portable terminal is held by the user's hand, it is difficult to completely fix the position of the portable terminal. Therefore, the background image is blurred.
 本実施形態によれば、ユーザによる選択操作で選択された画像ファイルの画像が背景画像(被重畳画像)として用いられるので、ユーザは、手振れを気にすることなく、かつ、楽な姿勢でAR画像を見ることができる。 According to the present embodiment, since the image of the image file selected by the selection operation by the user is used as the background image (superimposed image), the user does not mind camera shake and has an easy posture. You can see the image.
 また、ARコンテンツの視認画像は、その背景画像から検出されたマーカに基づいて設定される、3次元座標空間上の位置に置かれたARコンテンツが、そのマーカに基づいて算出される、背景画像の背景撮影位置及び背景撮影方向から視認される視界を表す。よって、ユーザは、表示されるAR画像を見ることにより、所望の背景画像に写る現実世界にあたかもそのARコンテンツが存在していたかのような状況を認識することができる。 The visual image of the AR content is set based on the marker detected from the background image, and the AR image placed at the position in the three-dimensional coordinate space is calculated based on the marker. Represents the field of view visually recognized from the background photographing position and the background photographing direction. Therefore, the user can recognize the situation as if the AR content existed in the real world shown in the desired background image by viewing the displayed AR image.
 以下、上述の第1の実施形態について更に詳細を説明する。以下には、詳細な実施形態として第2の実施形態及び第3の実施形態を例示する。以下では、上述の第1の実施形態と同じ内容については適宜省略する。以下では、上述の第1の実施形態と異なる内容を中心に説明する。 Hereinafter, further details of the above-described first embodiment will be described. Below, 2nd Embodiment and 3rd Embodiment are illustrated as detailed embodiment. Hereinafter, the same contents as those in the first embodiment are appropriately omitted. The following description will focus on the content different from the first embodiment described above.
[第2の実施形態]
 第2の実施形態における画像処理装置は、現実オブジェクトを複数の方向から撮影して得られた複数のコンテンツ画像の中の1つのコンテンツ画像を、ARコンテンツの視認画像として用いて、AR画像(合成画像)を生成する。
[Second Embodiment]
The image processing apparatus according to the second embodiment uses one content image among a plurality of content images obtained by photographing a real object from a plurality of directions as a visual image of the AR content, and uses an AR image (composite). Image).
 〔装置構成〕
 図3は、第2の実施形態における画像処理装置10のハードウェア構成例を概念的に示す図である。画像処理装置10は、いわゆるコンピュータである。携帯型、据え置き型等、画像処理装置10の形態は制限されない。画像処理装置10は、例えば、バスで相互に接続される、CPU(Central Processing Unit)11、メモリ12、入出力インタフェース(I/F)13、及び通信ユニット14等を有する。メモリ12は、RAM(Random Access Memory)、ROM(Read Only Memory)、及びハードディスク等の少なくともいずれかである。通信ユニット14は、他のコンピュータや機器と信号のやりとりを行う。通信ユニット14には、可搬型記録媒体18等も接続され得る。
〔Device configuration〕
FIG. 3 is a diagram conceptually illustrating a hardware configuration example of the image processing apparatus 10 according to the second embodiment. The image processing apparatus 10 is a so-called computer. The form of the image processing apparatus 10 such as a portable type or a stationary type is not limited. The image processing apparatus 10 includes, for example, a CPU (Central Processing Unit) 11, a memory 12, an input / output interface (I / F) 13, a communication unit 14, and the like that are connected to each other via a bus. The memory 12 is at least one of a RAM (Random Access Memory), a ROM (Read Only Memory), and a hard disk. The communication unit 14 exchanges signals with other computers and devices. A portable recording medium 18 or the like can be connected to the communication unit 14.
 入出力I/F13は、表示装置15と、入力装置16等のユーザインタフェース装置と接続可能である。表示装置15は、例えば、LCD(Liquid Crystal Display)やCRT(Cathode Ray Tube)ディスプレイ等の、CPU11やGPU(Graphics Processing Unit)(図示せず)等により処理された描画データに基づいて生成された画面を表示する装置である。表示装置15は、上述の表示部108に相当する。入力装置16は、例えば、キーボード、及びマウス等の、ユーザ操作の入力を受け付ける装置である。表示装置15及び入力装置16は一体化されていてもよい。表示装置15及び入力装置16は、例えばタッチパネルとして実現されてもよい。また、表示装置15及び入力装置16は、画像処理装置10と通信ユニット14を介して通信可能に接続される、他のコンピュータに接続されていてもよい。 The input / output I / F 13 can be connected to the display device 15 and a user interface device such as the input device 16. The display device 15 is generated based on drawing data processed by a CPU 11 or a GPU (Graphics Processing Unit) (not shown) such as an LCD (Liquid Crystal Display) or a CRT (Cathode Ray Tube) display. It is a device that displays a screen. The display device 15 corresponds to the display unit 108 described above. The input device 16 is a device that receives an input of a user operation, such as a keyboard and a mouse. The display device 15 and the input device 16 may be integrated. The display device 15 and the input device 16 may be realized as a touch panel, for example. The display device 15 and the input device 16 may be connected to another computer that is communicably connected to the image processing device 10 via the communication unit 14.
 また、画像処理装置10は、図3には示されない他のハードウェア要素を有していてもよい。例えば、画像処理装置10は、撮像部に相当するカメラを有していてもよい。この場合、画像処理装置10は、そのカメラで撮像された背景画像の画像ファイルを、背景保持部31に保持させることができる。そして、画像処理装置10は、後述する現実オブジェクトのコンテンツ画像を、そのカメラで撮像することができる。画像処理装置10のハードウェア構成は制限されない。 Further, the image processing apparatus 10 may have other hardware elements not shown in FIG. For example, the image processing apparatus 10 may have a camera corresponding to the imaging unit. In this case, the image processing apparatus 10 can cause the background holding unit 31 to hold the image file of the background image captured by the camera. Then, the image processing apparatus 10 can capture a content image of a real object described later with the camera. The hardware configuration of the image processing apparatus 10 is not limited.
 〔処理構成〕
 図4は、第2の実施形態における画像処理装置10の処理構成例を概念的に示す図である。第2の実施形態における画像処理装置10は、操作受付部21、背景取得部22、マーカ検出部23、基準算出部24、位置決定部25、画像抽出部26、サイズ調整部27、合成処理部28、及び表示制御部29等を有する。画像処理装置10は、さらに、背景保持部31、及びコンテンツ保持部32等を有する。これらの処理部は、例えば、CPU11によりメモリ12に格納されるプログラムが実行されることにより実現される。また、当該プログラムは、例えば、CD(Compact Disc)、メモリカード等のような可搬型記録媒体18やネットワーク上の他のコンピュータから入出力I/F13又は通信ユニット14を介してインストールされ、メモリ12に格納されてもよい。
[Processing configuration]
FIG. 4 is a diagram conceptually illustrating a processing configuration example of the image processing apparatus 10 according to the second embodiment. The image processing apparatus 10 according to the second embodiment includes an operation reception unit 21, a background acquisition unit 22, a marker detection unit 23, a reference calculation unit 24, a position determination unit 25, an image extraction unit 26, a size adjustment unit 27, and a synthesis processing unit. 28, a display control unit 29, and the like. The image processing apparatus 10 further includes a background holding unit 31, a content holding unit 32, and the like. These processing units are realized, for example, by executing a program stored in the memory 12 by the CPU 11. In addition, the program is installed from a portable recording medium 18 such as a CD (Compact Disc) or a memory card or another computer on the network via the input / output I / F 13 or the communication unit 14, and the memory 12. May be stored.
 操作受付部21は、上述の操作受付部101に相当する。操作受付部21は、表示装置15に表示される画面に対する入力装置16を用いたユーザ操作を受け付ける。操作受付部21は、画像ファイルの選択操作に加えて、合成画像に対するコンテンツ画像の移動操作を更に受け付ける。当該選択操作は、例えば、表示装置15に表示される複数の画像ファイルのリストから1つの画像ファイルを選択する操作と、その操作を確定するボタン操作とから構成される。また、その移動操作は、例えば、表示装置15に表示される合成画像上のコンテンツ画像をドラッグアンドドロップ又はスワイプする操作である。但し、本実施形態は、具体的な操作内容自体を制限しない。コンテンツ画像は、背景画像にARコンテンツの視認画像として重畳される画像である。コンテンツ画像について、後で詳細に説明する。 The operation reception unit 21 corresponds to the operation reception unit 101 described above. The operation reception unit 21 receives a user operation using the input device 16 for a screen displayed on the display device 15. The operation accepting unit 21 further accepts a content image moving operation with respect to the composite image in addition to the image file selecting operation. The selection operation includes, for example, an operation for selecting one image file from a list of a plurality of image files displayed on the display device 15 and a button operation for confirming the operation. The movement operation is, for example, an operation of dragging and dropping or swiping the content image on the composite image displayed on the display device 15. However, the present embodiment does not limit the specific operation content itself. The content image is an image that is superimposed as a visual image of the AR content on the background image. The content image will be described in detail later.
 更に、操作受付部21は、複数の現実オブジェクトの中から表示対象とする1つの現実オブジェクトを選択する操作を受け付けることもできる。現実オブジェクトは、背景画像に重畳されるARコンテンツの元となる、現実世界に存在する物や人である。操作受付部21は、後述するコンテンツ保持部32に保持されるコンテンツ画像に関連する複数の現実オブジェクトのリストを表示装置15に表示させることができる。そして、操作受付部21は、そのリストに対する入力装置16を用いた選択操作を受け付けることができる。 Furthermore, the operation accepting unit 21 can accept an operation for selecting one real object to be displayed from among a plurality of real objects. A real object is an object or person existing in the real world that is the source of AR content superimposed on a background image. The operation accepting unit 21 can cause the display device 15 to display a list of a plurality of real objects related to content images held in the content holding unit 32 described later. And the operation reception part 21 can receive selection operation using the input device 16 with respect to the list | wrist.
 背景取得部22、背景保持部31、マーカ検出部23及び基準算出部24は、上述の第1の実施形態における、背景取得部102、背景保持部107、マーカ検出部103及び基準算出部104と同様である。 The background acquisition unit 22, the background holding unit 31, the marker detection unit 23, and the reference calculation unit 24 are the same as the background acquisition unit 102, the background holding unit 107, the marker detection unit 103, and the reference calculation unit 104 in the first embodiment described above. It is the same.
 位置決定部25は、上述の位置決定部105のように、表示対象のARコンテンツを仮想的に配置する、3次元座標空間上の位置を決定する。更に、位置決定部25は、操作受付部21によって移動操作が受け付けられた場合に、その移動操作が示す、3次元座標空間上における、ARコンテンツの移動後の位置を決定する。具体的には、位置決定部25は、マーカの現実情報及び画像情報に基づいて、合成画像上における、移動操作が示す移動後の位置を、3次元座標空間上の位置に変換する。位置決定部25は、その変換された位置を、ARコンテンツの移動後の位置に決定する。 The position determination unit 25 determines the position in the three-dimensional coordinate space where the AR content to be displayed is virtually arranged, like the position determination unit 105 described above. Furthermore, when the movement operation is received by the operation reception unit 21, the position determination unit 25 determines a position after movement of the AR content on the three-dimensional coordinate space indicated by the movement operation. Specifically, the position determination unit 25 converts the moved position indicated by the moving operation on the composite image into a position in the three-dimensional coordinate space based on the marker actual information and image information. The position determination unit 25 determines the converted position as the position after movement of the AR content.
 コンテンツ保持部32は、ARコンテンツに相当する、現実オブジェクトを複数の方向から撮影することによって得られた、それぞれコンテンツ撮影方向に関連付けられた複数のコンテンツ画像を保持する。コンテンツ撮影方向は、コンテンツ画像が撮影された方向である。第2の実施形態では、当該複数のコンテンツ画像がARコンテンツの視認画像として用いられるので、コンテンツ保持部32は、細かい角度幅で全方向から撮影された多数のコンテンツ画像を保持することが望ましい。その理由は、背景画像に重畳するコンテンツ画像を、背景画像の背景撮影方向及び背景撮影位置からの現実オブジェクトの視認画像に近似させることができるからである。 The content holding unit 32 holds a plurality of content images respectively associated with the content shooting directions obtained by shooting a real object corresponding to the AR content from a plurality of directions. The content shooting direction is the direction in which the content image is shot. In the second embodiment, since the plurality of content images are used as visually recognized images of the AR content, it is desirable that the content holding unit 32 hold a large number of content images taken from all directions with a fine angular width. The reason is that the content image superimposed on the background image can be approximated to the visual image of the real object from the background shooting direction and the background shooting position of the background image.
 図5は、コンテンツ保持部32に保持される、さまざまなコンテンツ撮影方向に関連付けられた複数のコンテンツ画像の例を概念的に示す図である。図5では、コンテンツ撮影方向が、水平方向と垂直方向との2軸の座標空間に並べられて表記されている。図5に示されるように、コンテンツ保持部32は、それぞれ、コンテンツ撮影方向を示す位置座標と関連付けられた、複数のコンテンツ画像を保持することもできる。 FIG. 5 is a diagram conceptually showing an example of a plurality of content images held in the content holding unit 32 and associated with various content shooting directions. In FIG. 5, the content shooting direction is shown in a two-axis coordinate space in the horizontal direction and the vertical direction. As shown in FIG. 5, the content holding unit 32 can also hold a plurality of content images each associated with a position coordinate indicating the content shooting direction.
 また、図5の例では、各コンテンツ画像には、現実オブジェクトと共にマーカ及びその他の背景が写っている。第2の実施形態では、コンテンツ保持部32は、現実オブジェクト以外のものが写るコンテンツ画像を保持してもよい。但し、コンテンツ保持部32は、現実オブジェクト以外の画像領域が透明化されたコンテンツ画像を保持することが望ましい。その理由は、コンテンツ画像として、現実オブジェクトの視認画像のみを、背景画像に重畳させた合成画像を生成し易いからである。 In the example of FIG. 5, each content image includes a marker and other backgrounds along with the real object. In the second embodiment, the content holding unit 32 may hold a content image in which something other than a real object is captured. However, it is desirable that the content holding unit 32 holds a content image in which an image area other than the real object is made transparent. This is because, as a content image, it is easy to generate a composite image in which only a visual image of a real object is superimposed on a background image.
 コンテンツ保持部32は、保持するコンテンツ画像にマーカが写る場合には、そのマーカの現実世界のサイズ及び形状の情報を保持する。コンテンツ保持部32は、保持するコンテンツ画像にマーカが写らない場合には、現実オブジェクトの現実世界のサイズの情報を保持する。 The content holding unit 32 holds information on the size and shape of the marker in the real world when a marker appears in the content image to be held. The content holding unit 32 holds information about the size of the real world of the real object when the marker is not captured in the content image to be held.
 また、コンテンツ保持部32は、複数の現実オブジェクトの各々について、さまざまなコンテンツ撮影方向に関連付けられている複数のコンテンツ画像を保持することもできる。この場合、操作受付部21は、複数の現実オブジェクトの中から、表示対象とする1つの現実オブジェクトを選択する操作を受け付けることができる。 The content holding unit 32 can also hold a plurality of content images associated with various content shooting directions for each of the plurality of real objects. In this case, the operation reception unit 21 can receive an operation of selecting one real object to be displayed from among a plurality of real objects.
 画像抽出部26は、コンテンツ保持部32から、位置決定部25によって決定された位置から基準算出部24で算出された背景撮影位置への方向に近いコンテンツ撮影方向に関連付けられて保持されるコンテンツ画像を抽出する。ARコンテンツの3次元座標空間における位置から背景撮影位置への方向と一致するコンテンツ撮影方向に関連付けられているコンテンツ画像が、コンテンツ保持部32に保持されていることが望ましい。その理由は、そのコンテンツ画像が、背景画像の撮影位置からの現実オブジェクトの視認画像に一致するからである。そのように一致するコンテンツ画像がコンテンツ保持部32に保持されていない場合に、画像抽出部26は、最も近似するコンテンツ撮影方向に関連付けられて保持されるコンテンツ画像を抽出する。 The image extraction unit 26 holds the content image associated with the content shooting direction close to the direction from the position determined by the position determination unit 25 to the background shooting position calculated by the reference calculation unit 24 from the content holding unit 32. To extract. It is desirable that the content image associated with the content shooting direction that coincides with the direction from the position of the AR content in the three-dimensional coordinate space to the background shooting position is held in the content holding unit 32. The reason is that the content image matches the visual image of the real object from the shooting position of the background image. When such a matching content image is not held in the content holding unit 32, the image extraction unit 26 extracts a content image held in association with the closest content shooting direction.
 コンテンツ保持部32に複数の現実オブジェクトのコンテンツ画像が保持されている場合には、操作受付部21は、それらの複数の現実オブジェクトから1つの現実オブジェクトを選択する操作を受け付ける。画像抽出部26は、コンテンツ保持部32に保持される、選択された現実オブジェクトに関連付けられている複数のコンテンツ画像の中から、1つのコンテンツ画像を抽出する。 When content images of a plurality of real objects are held in the content holding unit 32, the operation receiving unit 21 receives an operation of selecting one real object from the plurality of real objects. The image extraction unit 26 extracts one content image from a plurality of content images associated with the selected real object that is retained in the content retention unit 32.
 また、操作受付部21によって移動操作が受け付けられた場合には、位置決定部25は移動後の位置を決定する。画像抽出部26は、位置決定部25により決定された移動後の位置から基準算出部24で算出された背景撮影位置への方向と近いコンテンツ撮影方向に対応付けられてコンテンツ保持部32に保持されるコンテンツ画像を新たに抽出する。新たに抽出されるコンテンツ画像は、背景画像の撮影位置からの移動後の位置に存在する現実オブジェクトの視認画像となる。 Further, when a movement operation is received by the operation reception unit 21, the position determination unit 25 determines a position after the movement. The image extraction unit 26 is held in the content holding unit 32 in association with a content shooting direction that is close to the direction from the moved position determined by the position determination unit 25 to the background shooting position calculated by the reference calculation unit 24. New content images are extracted. The newly extracted content image is a visual image of a real object that exists at a position after the background image is moved from the shooting position.
 サイズ調整部27は、現実オブジェクトの現実サイズと、マーカ検出部23で検出されたマーカに関する現実情報及び画像情報と、位置決定部25により決定される位置とに基づいて、画像抽出部26により抽出されたコンテンツ画像を拡大又は縮小する。サイズ調整部27は、コンテンツ画像に写る現実オブジェクトが現実世界のサイズで背景画像に写る現実世界と重畳されるように、コンテンツ画像の拡大率又は縮小率を決定する。例えば、サイズ調整部27は、マーカに関する現実情報及び画像情報を用いることによって、背景画像における現実世界の縮小率を得ることができる。サイズ調整部27は、現実オブジェクトの現実サイズと、コンテンツ画像での画像サイズと、背景画像における現実世界の縮小率と、背景撮影位置とARコンテンツの位置との間の距離とを用いて、コンテンツ画像の拡大率又は縮小率を決定することができる。 The size adjustment unit 27 is extracted by the image extraction unit 26 based on the actual size of the real object, the real information and image information regarding the marker detected by the marker detection unit 23, and the position determined by the position determination unit 25. The enlarged content image is enlarged or reduced. The size adjusting unit 27 determines the enlargement ratio or reduction ratio of the content image so that the real object shown in the content image is superimposed on the real world shown in the background image in the size of the real world. For example, the size adjusting unit 27 can obtain the reduction ratio of the real world in the background image by using the real information and the image information regarding the marker. The size adjustment unit 27 uses the real size of the real object, the image size in the content image, the reduction ratio of the real world in the background image, and the distance between the background shooting position and the AR content position. The enlargement or reduction ratio of the image can be determined.
 合成処理部28は、3次元座標空間における、位置決定部25により決定された位置に配置されるARコンテンツの、背景撮影位置及び背景撮影方向からの視認画像が、背景画像に重畳された合成画像(AR画像)を生成する。具体的には、合成処理部28は、画像抽出部26により抽出されたコンテンツ画像を、現実オブジェクト(ARコンテンツ)の視認画像として、背景取得部22により取得された背景画像に重畳する。 The composition processing unit 28 is a composite image in which the visual image from the background photographing position and the background photographing direction of the AR content arranged at the position determined by the position determining unit 25 in the three-dimensional coordinate space is superimposed on the background image. (AR image) is generated. Specifically, the composition processing unit 28 superimposes the content image extracted by the image extraction unit 26 on the background image acquired by the background acquisition unit 22 as a visual image of the real object (AR content).
 合成処理部28は、背景画像に重畳するコンテンツ画像を、操作受付部21による移動操作の受け付けに応じて画像抽出部26により新たに抽出されたコンテンツ画像に切り換える。また、サイズ調整部27によりコンテンツ画像が拡大又は縮小された場合には、合成処理部28は、その拡大又は縮小されたコンテンツ画像を背景画像に重畳する。 The composition processing unit 28 switches the content image to be superimposed on the background image to the content image newly extracted by the image extraction unit 26 in response to the movement operation received by the operation reception unit 21. Further, when the content image is enlarged or reduced by the size adjusting unit 27, the composition processing unit 28 superimposes the enlarged or reduced content image on the background image.
 表示制御部29は、合成処理部28により生成された合成画像を表示装置15に表示させる。 The display control unit 29 causes the display device 15 to display the composite image generated by the composition processing unit 28.
 〔動作例〕
 以下、第2の実施形態における画像処理方法について図6を用いて説明する。図6は、第2の実施形態における画像処理装置10の動作例を示すフローチャートである。以下の説明では、画像処理装置10が各処理工程(すなわちステップ)を実行する。しかし、画像処理装置10に含まれる上述のいずれかの処理部のうち、実行される処理工程に応じた処理部が、その処理工程を実行してもよい。図6では、図2に示される処理工程と同じ処理工程には、図2と同じ符号が付されている。
[Operation example]
The image processing method according to the second embodiment will be described below with reference to FIG. FIG. 6 is a flowchart illustrating an operation example of the image processing apparatus 10 according to the second embodiment. In the following description, the image processing apparatus 10 executes each processing step (that is, step). However, among any of the above-described processing units included in the image processing apparatus 10, a processing unit corresponding to the processing step to be executed may execute the processing step. 6, the same processing steps as those shown in FIG. 2 are denoted by the same reference numerals as those in FIG.
 画像処理装置10は、入力装置16を用いたユーザ操作を受け付ける(ステップS61)。ステップS61の内容は、操作受付部21の処理内容と同様である。
 画像処理装置10は、選択操作が受け付けられた場合には(ステップS62において「選択」)、図2と同様に、ステップS22からステップS26までの処理工程を実行する。
The image processing apparatus 10 accepts a user operation using the input device 16 (step S61). The content of step S61 is the same as the processing content of the operation reception part 21. FIG.
When the selection operation is accepted (“selection” in step S62), the image processing apparatus 10 executes the processing steps from step S22 to step S26 as in FIG.
 この場合、画像処理装置10は、コンテンツ保持部32に保持されるコンテンツ画像から、コンテンツ画像を抽出する(ステップS63)。ステップS63において、画像処理装置10は、ステップS26において決定された位置からステップS25において算出された背景撮影位置への方向と近いコンテンツ撮影方向に関連付けられているコンテンツ画像を抽出すればよい。ステップS63の内容は、画像抽出部26の処理内容と同様である。 In this case, the image processing apparatus 10 extracts a content image from the content image held in the content holding unit 32 (step S63). In step S63, the image processing apparatus 10 may extract a content image associated with a content shooting direction that is close to the direction from the position determined in step S26 to the background shooting position calculated in step S25. The content of step S63 is the same as the processing content of the image extraction unit 26.
 続いて、画像処理装置10は、現実オブジェクトが現実世界におけるサイズで背景画像に写る現実世界に重畳されるように、ステップS63で抽出されたコンテンツ画像を拡大又は縮小する(ステップS64)。ステップS64の内容は、サイズ調整部27の処理内容と同様である。 Subsequently, the image processing apparatus 10 enlarges or reduces the content image extracted in step S63 so that the real object is superimposed on the real world in the size of the real world and reflected in the background image (step S64). The content of step S64 is the same as the processing content of the size adjustment unit 27.
 画像処理装置10は、ステップS64において拡大又は縮小されたコンテンツ画像を、ステップS22において取得された背景画像に重畳することによって、合成画像を生成する(ステップS65)。ステップS65の内容は、合成処理部28の処理内容と同様である。 The image processing apparatus 10 generates a composite image by superimposing the content image enlarged or reduced in step S64 on the background image acquired in step S22 (step S65). The content of step S65 is the same as the processing content of the composition processing unit 28.
 画像処理装置10は、ステップS65において生成された合成画像を、表示装置15に表示させる(ステップS66)。 The image processing device 10 causes the display device 15 to display the composite image generated in step S65 (step S66).
 画像処理装置10は、移動操作が受け付けられた場合には(ステップS62において「移動」)、ステップS61において受け付けられた移動操作が示す、3次元座標空間上におけるARコンテンツの移動後の位置を決定する(ステップS67)。ステップS67の内容は、位置決定部25の処理内容と同様である。 When the moving operation is accepted (“move” in step S62), the image processing apparatus 10 determines the position after the movement of the AR content in the three-dimensional coordinate space indicated by the moving operation accepted in step S61. (Step S67). The content of step S67 is the same as the processing content of the position determination unit 25.
 画像処理装置10は、コンテンツ保持部32に保持されるコンテンツ画像から、コンテンツ画像を新たに抽出する(ステップS63)。この場合、画像処理装置10は、ステップS67において決定された移動後の位置からステップS25で算出された背景撮影位置への方向と近いコンテンツ撮影方向に関連付けられているコンテンツ画像を抽出すればよい。以降、画像処理装置10は、この新たに抽出されたコンテンツ画像に対して、ステップS64以降を実行する。 The image processing apparatus 10 newly extracts a content image from the content image held in the content holding unit 32 (step S63). In this case, the image processing apparatus 10 may extract a content image associated with a content shooting direction that is close to the direction from the moved position determined in step S67 to the background shooting position calculated in step S25. Thereafter, the image processing apparatus 10 executes step S64 and subsequent steps on the newly extracted content image.
 また、コンテンツ保持部32に複数の現実オブジェクトのコンテンツ画像が保持されている場合、画像処理装置10は、ステップS61において、複数の現実オブジェクトの中から表示対象とする1つの現実オブジェクトを選択する操作を受け付けることもできる。この場合、画像処理装置10は、ステップS63において、コンテンツ保持部32に保持される複数のコンテンツ画像の中から、その操作により選択された現実オブジェクトに関連付けられている、1つのコンテンツ画像を抽出する。以上の、コンテンツ保持部32に複数の現実オブジェクトのコンテンツ画像が保持されている場合における動作は、図6には図示されていない。 When content images of a plurality of real objects are held in the content holding unit 32, the image processing apparatus 10 performs an operation of selecting one real object to be displayed from the plurality of real objects in step S61. Can also be accepted. In this case, in step S63, the image processing apparatus 10 extracts one content image associated with the real object selected by the operation from the plurality of content images held in the content holding unit 32. . The above operation in the case where content images of a plurality of real objects are held in the content holding unit 32 is not shown in FIG.
 〔第2の実施形態の作用及び効果〕
 上述したように第2の実施形態では、複数の画像ファイルの中からユーザによる選択操作で選択された画像ファイルの画像が背景画像(被重畳画像)として用いられるので、ユーザに、場所及び時間の制限なく、所望の背景画像に、ARコンテンツが重畳されたAR画像を見せることができる。
[Operation and Effect of Second Embodiment]
As described above, in the second embodiment, an image of an image file selected by a user's selection operation from among a plurality of image files is used as a background image (superimposed image). Without limitation, an AR image in which AR content is superimposed on a desired background image can be shown.
 更に、第2の実施形態では、現実オブジェクトを複数の方向から撮影して得られた、それぞれコンテンツ撮影方向に関連付けられている複数のコンテンツ画像が、予め保持されている。そして、保持される複数のコンテンツ画像の中の1つのコンテンツ画像をARコンテンツの視認画像として用いて、AR画像(合成画像)が生成される。このように、第2の実施形態によれば、コンテンツ撮影方向と関連付けられた複数のコンテンツ画像を準備すればよいので、専門的なCG技術が必要となる3次元モデルを生成する形態に較べ、ARコンテンツを容易に作成することができる。従って、第2の実施形態によれば、AR技術を用いるプラットフォームの開発及び利用の敷居を低くすることができる。さらに、その、AR技術を用いるプラットフォームを一般ユーザに利用し易くすることができる。 Furthermore, in the second embodiment, a plurality of content images obtained by photographing a real object from a plurality of directions and respectively associated with the content photographing directions are stored in advance. Then, an AR image (composite image) is generated using one content image among the plurality of held content images as a visual image of the AR content. As described above, according to the second embodiment, since it is only necessary to prepare a plurality of content images associated with the content shooting direction, compared to a mode of generating a three-dimensional model that requires specialized CG technology, AR content can be easily created. Therefore, according to the second embodiment, the threshold for development and use of a platform using the AR technology can be reduced. Furthermore, the platform using the AR technology can be easily used by general users.
 更に、第2の実施形態では、3次元座標空間上のARコンテンツの位置から背景画像の撮影位置への方向と近いコンテンツ撮影方向に関連付けられている、保持されるコンテンツ画像が、現実オブジェクトの視認画像として背景画像に重畳される。よって、第2の実施形態によれば、背景画像に重畳するコンテンツ画像を、背景画像の撮影位置及び撮影方向からの現実オブジェクトの視認画像に近似させることができる。そのため、背景画像に写る現実世界に現実オブジェクトが重畳された空間の現実感を増強させることができる。 Furthermore, in the second embodiment, the held content image associated with the content shooting direction close to the direction from the AR content position to the background image shooting position in the three-dimensional coordinate space is the visual recognition of the real object. The image is superimposed on the background image. Therefore, according to the second embodiment, the content image to be superimposed on the background image can be approximated to the visual image of the real object from the shooting position and shooting direction of the background image. Therefore, it is possible to enhance the realism of the space in which the real object is superimposed on the real world shown in the background image.
 更に、第2の実施形態では、移動操作が示す3次元座標空間上の移動後の位置が決定される。さらに、決定された移動後の位置から背景画像の撮影位置への方向と近いコンテンツ撮影方向に関連付けられている、保持されるコンテンツ画像が新たに抽出される。背景画像に重畳するコンテンツ画像は、その新たに抽出されたコンテンツ画像に切り換えられる。よって、AR画像上でコンテンツ画像に対して移動操作が行われた場合、その移動後の位置に応じて、AR画像上における現実オブジェクトの見え方が変わる。このように、第2の実施形態によれば、現実感を維持しつつ、AR画像上でのARコンテンツとしての現実オブジェクトの、ユーザによる移動を可能にする。 Furthermore, in the second embodiment, the position after movement in the three-dimensional coordinate space indicated by the movement operation is determined. Furthermore, a content image to be held that is associated with a content shooting direction that is close to the direction from the determined position after movement to the shooting position of the background image is newly extracted. The content image to be superimposed on the background image is switched to the newly extracted content image. Therefore, when a movement operation is performed on the content image on the AR image, the appearance of the real object on the AR image changes according to the position after the movement. As described above, according to the second embodiment, the user can move the real object as the AR content on the AR image while maintaining the reality.
 更に、第2の実施形態では、現実オブジェクトが、現実世界でのサイズで、背景画像に写る現実世界に重畳されるように、重畳するコンテンツ画像が拡大又は縮小される。従って、第2の実施形態によれば、現実オブジェクトと背景画像に写る背景との現実世界におけるサイズ関係を、ユーザに適切に視認させることができる。 Furthermore, in the second embodiment, the content image to be superimposed is enlarged or reduced so that the real object is superimposed on the real world in the background image with the size in the real world. Therefore, according to the second embodiment, the size relationship in the real world between the real object and the background reflected in the background image can be appropriately recognized by the user.
 また、第2の実施形態では、複数の現実オブジェクトの各々について、さまざまなコンテンツ撮影方向に関連付けられる複数のコンテンツ画像を保持することができる。その場合、ユーザに、所望の背景画像のみでなく、重畳するARコンテンツとしての、所望の現実オブジェクトを、選択させることができる。 In the second embodiment, a plurality of content images associated with various content shooting directions can be held for each of the plurality of real objects. In this case, the user can select not only a desired background image but also a desired real object as an AR content to be superimposed.
[第3の実施形態]
 第3の実施形態における画像処理装置は、画像ファイルの画像を背景画像として取得する機能と、リアルタイム映像の画像フレームを背景画像として取得する機能とを選択的に動作させる。以下、第3の実施形態における画像処理装置10について、第2の実施形態と異なる内容を中心に説明する。以下の説明では、第2の実施形態と同様の内容については適宜省略する。
[Third Embodiment]
The image processing apparatus according to the third embodiment selectively operates a function of acquiring an image of an image file as a background image and a function of acquiring an image frame of a real-time video as a background image. Hereinafter, the image processing apparatus 10 according to the third embodiment will be described focusing on contents different from those of the second embodiment. In the following description, the same contents as those in the second embodiment are omitted as appropriate.
 〔装置構成〕
 図7は、第3の実施形態における画像処理装置10のハードウェア構成例を概念的に示す図である。第3の実施形態における画像処理装置10のハードウェア構成は、第2の実施形態の画像処理装置10のハードウェア構成と同じでよい。第3の実施形態における画像処理装置10は、第2の実施形態の構成に加えて、撮像部に相当するカメラ17を更に有する。すなわち、本実施形態における画像処理装置10には、更に、カメラ17が接続されている。カメラ17は、入出力I/F13又は通信ユニット14に接続される。
〔Device configuration〕
FIG. 7 is a diagram conceptually illustrating a hardware configuration example of the image processing apparatus 10 according to the third embodiment. The hardware configuration of the image processing apparatus 10 according to the third embodiment may be the same as the hardware configuration of the image processing apparatus 10 according to the second embodiment. The image processing apparatus 10 according to the third embodiment further includes a camera 17 that corresponds to an imaging unit in addition to the configuration of the second embodiment. That is, the camera 17 is further connected to the image processing apparatus 10 in the present embodiment. The camera 17 is connected to the input / output I / F 13 or the communication unit 14.
 〔処理構成〕
 第3の実施形態における画像処理装置10は、図4に示される第2の実施形態と同様の機能的な構成を有する。
[Processing configuration]
The image processing apparatus 10 according to the third embodiment has the same functional configuration as that of the second embodiment shown in FIG.
 操作受付部21は、カメラ17からのリアルタイム映像の取得を要求する操作(以下、要求操作)も受け付けることができる。この操作の受け付けに応じて、操作受付部21は、カメラ17を起動させることもできる。 The operation accepting unit 21 can also accept an operation for requesting acquisition of a real-time video from the camera 17 (hereinafter, a request operation). In response to accepting this operation, the operation accepting unit 21 can also activate the camera 17.
 背景取得部22は、操作受付部21によって上記要求操作が受け付けられた場合、その受け付けられた上記要求操作に応じて、カメラ17から得られるリアルタイム映像の画像フレームを、背景画像として逐次取得する。これにより、被重畳画像としての背景画像が画像ファイルの画像からリアルタイム映像の画像フレームに切り換えられる。 When the operation request is received by the operation reception unit 21, the background acquisition unit 22 sequentially acquires the real-time video image frame obtained from the camera 17 as a background image according to the received request operation. Thereby, the background image as the superimposed image is switched from the image of the image file to the image frame of the real-time video.
 表示制御部29は、上述のようなコンテンツ画像が視認画像として重畳されたリアルタイム映像を表示装置15に表示させる。 The display control unit 29 causes the display device 15 to display a real-time video in which the content image as described above is superimposed as a visual image.
 〔動作例〕
 以下、第3の実施形態における画像処理方法について図8を用いて説明する。図8は、第3の実施形態における画像処理装置10の動作例の一部を示すフローチャートである。以下の説明では、画像処理装置10が各処理工程を実行する。しかし、画像処理装置10に含まれる上述の処理部のうち、実行される処理工程に応じた処理部が、その処理工程を実行してもよい。図8では、図6に示される処理工程と同じ処理工程には、図6と同じ符号が付されている。
[Operation example]
Hereinafter, an image processing method according to the third embodiment will be described with reference to FIG. FIG. 8 is a flowchart illustrating a part of an operation example of the image processing apparatus 10 according to the third embodiment. In the following description, the image processing apparatus 10 executes each processing process. However, among the above-described processing units included in the image processing apparatus 10, a processing unit corresponding to the processing step to be executed may execute the processing step. In FIG. 8, the same processing steps as those shown in FIG. 6 are denoted by the same reference numerals as in FIG.
 第3の実施形態における画像処理装置10は、まず、ステップS61を実行する。画像処理装置10は、ステップS61において、リアルタイム映像の取得を要求する操作を受け付けた場合、図8に示される動作を実行する。
 画像処理装置10は、カメラ17から得られるリアルタイム映像の画像フレームを背景画像として取得する(ステップS81)。以降、ステップS81において取得された背景画像に対して、上述の第2の実施形態と同様に、ステップS23以降の処理工程を実行する。
The image processing apparatus 10 according to the third embodiment first executes step S61. When the image processing apparatus 10 receives an operation requesting acquisition of real-time video in step S61, the image processing apparatus 10 performs the operation shown in FIG.
The image processing apparatus 10 acquires an image frame of a real-time video obtained from the camera 17 as a background image (step S81). Thereafter, the processing steps after step S23 are executed on the background image acquired in step S81, as in the second embodiment described above.
 画像処理装置10は、カメラ17から得られるリアルタイム映像の画像フレーム毎に、図8に示される動作を繰り返し実行することによって、コンテンツ画像が視認画像として重畳されたリアルタイム映像を表示装置15に表示させることができる。画像処理装置10は、リアルタイム映像を構成する全画像フレームを背景画像として利用する必要はない。画像処理装置10は、例えば、リアルタイム映像を構成する画像フレームから、所定の方法に従って一部の画像フレームを除去する間引きを行ってもよい。画像処理装置10は、間引きが行われた後の画像フレームを、背景画像として利用してもよい。また、画像処理装置10は、隣接する画像フレーム間でマーカの配置が変わらない場合には、ステップS23からステップS64までの処理工程を省くことができる。 The image processing apparatus 10 displays the real-time video on which the content image is superimposed as a visual image on the display device 15 by repeatedly executing the operation shown in FIG. 8 for each image frame of the real-time video obtained from the camera 17. be able to. The image processing apparatus 10 does not have to use all the image frames constituting the real-time video as the background image. For example, the image processing apparatus 10 may perform thinning by removing some image frames from the image frames constituting the real-time video according to a predetermined method. The image processing apparatus 10 may use the image frame after thinning as a background image. Further, the image processing apparatus 10 can omit the processing steps from step S23 to step S64 when the marker arrangement does not change between adjacent image frames.
 〔第3の実施形態における作用及び効果〕
 第3の実施形態では、ユーザによる操作に応じて、ARコンテンツが重畳される背景画像が、画像ファイルの画像からリアルタイム映像の画像フレームに切り換えられる。そして、リアルタイム画像の画像フレームに対して、ARコンテンツの位置からその画像フレームの撮影位置への方向と近いコンテンツ撮影方向と関連付けられている、保持されるコンテンツ画像が、現実オブジェクトの視認画像として重畳される。これにより、第3の実施形態によれば、コンテンツ画像が視認画像として重畳されたリアルタイム映像を、表示装置15に表示させることができる。
[Operation and Effect in Third Embodiment]
In the third embodiment, the background image on which the AR content is superimposed is switched from the image of the image file to the image frame of the real-time video according to the operation by the user. Then, a held content image associated with a content shooting direction close to the direction from the AR content position to the shooting position of the image frame is superimposed on the image frame of the real-time image as a visual image of the real object. Is done. Thereby, according to 3rd Embodiment, the real-time image | video with which the content image was superimposed as a visual recognition image can be displayed on the display apparatus 15. FIG.
[第1の変形例]
 第1の変形例は、本発明の第2の実施形態の変形例である。本発明の第3の実施形態を、本変形例のように変形することも可能である。上述の第2の実施形態では、ユーザによる背景画像の選択操作が可能であるので、ユーザが、マーカが写っていない背景画像が記録されている画像ファイルを、誤って選択する可能性がある。コンピュータに格納される画像ファイルは多数存在する。そして、その中には、適切な背景画像を保存する画像ファイル以外の画像ファイルも多数存在する。そのため、このような誤りは起こり得る。そして、マーカが背景画像に写っていない場合、その背景画像に適切にARコンテンツを重畳することはできない。そこで、第2の実施形態における画像処理装置10の、以下の処理部は、次のような処理を行ってもよい。
[First Modification]
The first modification is a modification of the second embodiment of the present invention. The third embodiment of the present invention can be modified as in this modification. In the second embodiment described above, since the user can select a background image, the user may erroneously select an image file in which a background image without a marker is recorded. There are many image files stored in the computer. Among them, there are many image files other than an image file for storing an appropriate background image. Therefore, such an error can occur. If the marker is not reflected in the background image, the AR content cannot be appropriately superimposed on the background image. Therefore, the following processing unit of the image processing apparatus 10 according to the second embodiment may perform the following processing.
 表示制御部29は、マーカ検出部23により背景画像からマーカが検出されない場合に、マーカが検出されないことを表示装置15に表示させる。
 操作受付部21は、マーカが検出されなかった画像ファイル以外の画像ファイルを選択し直す操作を受け付ける。
 背景取得部22は、選択し直す操作により新たに選択された画像ファイルの画像を、背景画像として新たに取得する。
The display control unit 29 causes the display device 15 to display that no marker is detected when no marker is detected from the background image by the marker detection unit 23.
The operation accepting unit 21 accepts an operation of reselecting an image file other than the image file in which no marker is detected.
The background acquisition unit 22 newly acquires an image of the image file newly selected by the reselection operation as a background image.
 これにより、ユーザは、選択した画像ファイルに、ARコンテンツを重畳させることができない背景画像が保存されているか否かを認識することができる。さらに、ユーザは、画像ファイルを選択し直すことができる。そのため、ユーザは、安心して、画像ファイルの選択操作を行うことができる。 Thereby, the user can recognize whether or not a background image that cannot superimpose the AR content is stored in the selected image file. Furthermore, the user can reselect the image file. Therefore, the user can perform an image file selection operation with peace of mind.
[第2の変形例]
 第2の変形例は、本発明の第2の実施形態の変形例である。本発明の第3の実施形態を、本変形例のように変形することも可能である。背景画像を保存する画像ファイルを生成する時間及び場所は任意である。言い換えれば、ユーザは、異なる複数の場所を写した写真が記録されている複数の画像ファイルを、それらの複数の画像ファイルを背景画像として用いるために、背景保持部31に保持させることができる。このような状況下では、複数の種種のマーカを提供した方が、利便性が上がる可能性がある。一方、画像処理装置10は、背景画像に写るマーカの現実情報(形状情報、及びサイズ情報)を知っている必要がある。
[Second Modification]
The second modification is a modification of the second embodiment of the present invention. The third embodiment of the present invention can be modified as in this modification. The time and place for generating the image file for storing the background image are arbitrary. In other words, the user can hold a plurality of image files in which photographs showing a plurality of different places are recorded in the background holding unit 31 in order to use the plurality of image files as a background image. Under such circumstances, it may be more convenient to provide multiple types of markers. On the other hand, the image processing apparatus 10 needs to know the actual information (shape information and size information) of the marker that appears in the background image.
 図9は、第2の変形例における画像処理装置10Aの処理構成例を概念的に示す図である。第2の変形例における画像処理装置10Aは、第2の実施形態及び第3の実施形態の構成に加えて、マーカ情報保持部35を更に有する。マーカ情報保持部35についても他の処理部と同様に実現される。画像処理装置10Aではなく、画像処理装置10Aと通信可能に接続される他のコンピュータが、マーカ情報保持部35を有していてもよい。 FIG. 9 is a diagram conceptually illustrating a processing configuration example of the image processing apparatus 10A according to the second modification. The image processing apparatus 10A according to the second modification further includes a marker information holding unit 35 in addition to the configurations of the second and third embodiments. The marker information holding unit 35 is also realized in the same manner as other processing units. Instead of the image processing apparatus 10 </ b> A, another computer that is communicably connected to the image processing apparatus 10 </ b> A may have the marker information holding unit 35.
 マーカ情報保持部35は、複数の、マーカの種類の各々について、形状情報及びサイズ情報を含む、マーカの情報を保持する。この場合、マーカ検出部23は、マーカ情報保持部35から、特定した種類のマーカの形状情報及びサイズ情報を取得する。マーカ検出部23は、この取得された形状情報及びサイズ情報を用いて、背景取得部22で取得された背景画像からマーカを検出する。マーカの種類を特定する手法は制限されない。例えば、マーカ検出部23は、マーカ情報保持部35に保持されるマーカの種類毎に、順次、マーカが検出できるか否かを試すことができる。また、マーカ検出部23は、ユーザにマーカの種類を選択させるための画面を、表示装置15に表示させることができる。マーカ検出部23は、ユーザによる、マーカの種類を選択する操作に応じて、マーカの種類を特定することもできる。 The marker information holding unit 35 holds marker information including shape information and size information for each of a plurality of marker types. In this case, the marker detection unit 23 acquires shape information and size information of the specified type of marker from the marker information holding unit 35. The marker detection unit 23 detects a marker from the background image acquired by the background acquisition unit 22 using the acquired shape information and size information. The method for specifying the marker type is not limited. For example, the marker detection unit 23 can test whether or not a marker can be sequentially detected for each type of marker held in the marker information holding unit 35. In addition, the marker detection unit 23 can cause the display device 15 to display a screen for allowing the user to select a marker type. The marker detection unit 23 can also specify the marker type in response to an operation of selecting the marker type by the user.
 このようにすれば、背景画像を撮影するにあたり、ユーザに複数種のマーカを用いることを許容することができる。そして、ARプラットフォームの利便性を向上させることができる。 In this way, it is possible to allow the user to use a plurality of types of markers when taking a background image. The convenience of the AR platform can be improved.
[第3の変形例]
 第3の変形例は、本発明の第3の実施形態の変形例である。第3の実施形態の画像処理装置10の、以下の処理部は、次のような処理を行ってもよい。
[Third Modification]
The third modification is a modification of the third embodiment of the present invention. The following processing unit of the image processing apparatus 10 according to the third embodiment may perform the following processing.
 操作受付部21は、視認画像が重畳されたリアルタイム映像が表示装置15に表示されている状態において、スナップショット操作を受け付ける。
 背景取得部22は、操作受付部21によるスナップショット操作の受け付けに応じて、カメラ17によって得られるリアルタイム映像の画像フレームを、背景画像として固定的に保持する。このとき、背景取得部22は、画像フレームを固定的に保持した後、カメラ17を停止することもできる。
 表示制御部29は、視認画像が重畳されたリアルタイム映像の表示に替えて、固定的に保持される背景画像に、コンテンツ画像が視認画像として重畳された合成画像を、表示装置15に表示させる。
The operation reception unit 21 receives a snapshot operation in a state where a real-time video on which a visual image is superimposed is displayed on the display device 15.
The background acquisition unit 22 fixedly holds the image frame of the real-time video obtained by the camera 17 as the background image in response to reception of the snapshot operation by the operation reception unit 21. At this time, the background acquisition unit 22 can also stop the camera 17 after holding the image frame fixedly.
The display control unit 29 causes the display device 15 to display a composite image in which a content image is superimposed as a visual image on a fixedly held background image, instead of displaying a real-time video on which the visual image is superimposed.
 このようにすれば、ユーザによるスナップショット操作が受け付けられた際に得られた画像フレームを固定的に背景画像として用いて、コンテンツ画像が重畳されたAR画像を表示装置15に表示させることができる。これによれば、ユーザは、リアルタイム映像を好きなタイミングで停止させた画像に、コンテンツ画像が重畳されたAR画像を見ることができる。更に、これによれば、重畳されるコンテンツ画像の位置が手振れ等により揺れることによって、表示されるAR画像が見難くなることを防止することができ、かつ、ユーザに楽な姿勢でAR画像を見させることができる。 In this way, the AR image on which the content image is superimposed can be displayed on the display device 15 by using the image frame obtained when the snapshot operation by the user is accepted as a fixed background image. . According to this, the user can see the AR image in which the content image is superimposed on the image in which the real-time video is stopped at a desired timing. Further, according to this, it is possible to prevent the displayed AR image from becoming difficult to see due to the position of the superimposed content image being shaken by hand shake or the like, and the AR image can be displayed in an easy posture for the user. Can be seen.
[補足]
 第3の実施形態における画像処理装置10は、画像ファイルの画像を背景画像として取得する機能と、リアルタイム映像の画像フレームを背景画像として取得する機能とを選択的に動作させる。しかし、画像処理装置10は、画像ファイルの画像を背景画像として取得する機能を持たず、リアルタイム映像の画像フレームを背景画像として取得する機能を動作させてもよい。この場合、画像処理装置10は操作受付部21を持たなくてもよい。
[Supplement]
The image processing apparatus 10 according to the third embodiment selectively operates a function of acquiring an image of an image file as a background image and a function of acquiring an image frame of a real-time video as a background image. However, the image processing apparatus 10 may operate the function of acquiring the image frame of the real-time video as the background image without having the function of acquiring the image of the image file as the background image. In this case, the image processing apparatus 10 may not have the operation receiving unit 21.
 以下にあげる実装例によって、上述の各実施形態を更に詳細に説明する。本発明は以下の実装例によって何ら限定を受けない。
 [実装例]
The above-described embodiments will be described in more detail with the following implementation examples. The present invention is not limited by the following implementation examples.
[Example of implementation]
 図10A、図10B及び図10Cは、画像ファイルに各々保存される背景画像の例をそれぞれ示す図である。これら図に示されるように、各背景画像にはそれぞれマーカが含まれる。これら例では、同じマーカが各背景画像に含まれているが、画像処理装置10が知っているマーカであれば、各背景画像に含まれるマーカは同じでなくてもよい。これら例に示されるように、背景保持部31には、異なる各背景画像をそれぞれ保存する複数の画像ファイルが保持される。 FIG. 10A, FIG. 10B, and FIG. 10C are diagrams respectively showing examples of background images stored in the image file. As shown in these figures, each background image includes a marker. In these examples, the same marker is included in each background image, but the marker included in each background image may not be the same as long as it is a marker known by the image processing apparatus 10. As shown in these examples, the background holding unit 31 holds a plurality of image files for storing different background images.
 以下では、図6及び図8で示されるステップS63及びステップS65、並びに、画像抽出部26及び合成処理部28の処理内容の更なる具体例を説明する。この場合、コンテンツ保持部32は、コンテンツ画像を保持する。コンテンツ保持部32が保持する各コンテンツ画像には、コンテンツ撮影方向を表す情報として、現実オブジェクトの撮影時における現実オブジェクトからカメラ17への方向を表す単位ベクトルが関連付けられる。 Hereinafter, further specific examples of the processing contents of steps S63 and S65 shown in FIGS. 6 and 8 and the image extraction unit 26 and the composition processing unit 28 will be described. In this case, the content holding unit 32 holds a content image. Each content image held by the content holding unit 32 is associated with a unit vector representing the direction from the real object to the camera 17 when photographing the real object as information representing the content photographing direction.
 画像処理装置10(画像抽出部26)は、3次元座標空間上のARコンテンツの位置から背景画像の撮影位置への方向を示す単位ベクトルを算出する。画像処理装置10(画像抽出部26)は、背景画像から算出された単位ベクトルと、コンテンツ保持部32に保持される各コンテンツ画像の各単位ベクトルとの内積をそれぞれ計算する。画像処理装置10(画像抽出部26)は、計算された内積値が最大である、コンテンツ画像の単位ベクトルを特定する。画像処理装置10(画像抽出部26)は、その特定された単位ベクトルと関連付けられている、保持されているコンテンツ画像を、コンテンツ保持部32から抽出する。 The image processing apparatus 10 (image extraction unit 26) calculates a unit vector indicating a direction from the position of the AR content in the three-dimensional coordinate space to the shooting position of the background image. The image processing apparatus 10 (image extraction unit 26) calculates the inner product of the unit vector calculated from the background image and each unit vector of each content image held in the content holding unit 32. The image processing apparatus 10 (image extraction unit 26) specifies the unit vector of the content image that has the maximum calculated inner product value. The image processing apparatus 10 (image extraction unit 26) extracts the held content image associated with the specified unit vector from the content holding unit 32.
 画像処理装置10(合成処理部28)は、コンテンツ画像と背景画像とを合成する際に、まず、3次元座標空間上において定義される、透明な板状の3次元オブジェクトを生成しておいてもよい。この3次元オブジェクトは、仮想的なオブジェクトである。その3次元オブジェクトの画像が、背景画像に重畳される。画像処理装置10(合成処理部28)は、上述のように抽出されたコンテンツ画像を板状の3次元オブジェクトにテクスチャとして貼り付けてもよい。そのことにより、画像処理装置10(合成処理部28)は、背景画像にコンテンツ画像を重畳する。画像処理装置10(合成処理部28)は、コンテンツ画像が貼り付けられた3次元オブジェクトの像が、ほぼ、背景撮影位置における現実オブジェクトの視認画像であるように、3次元オブジェクトの位置及びコンテンツ画像を貼り付ける方法を設定すればよい。背景画像にコンテンツ画像を重畳する方法は、以上の例に限られない。 When synthesizing the content image and the background image, the image processing apparatus 10 (the synthesis processing unit 28) first generates a transparent plate-shaped three-dimensional object defined in the three-dimensional coordinate space. Also good. This three-dimensional object is a virtual object. The image of the three-dimensional object is superimposed on the background image. The image processing apparatus 10 (the composition processing unit 28) may paste the content image extracted as described above as a texture on a plate-shaped three-dimensional object. As a result, the image processing apparatus 10 (the composition processing unit 28) superimposes the content image on the background image. The image processing apparatus 10 (the composition processing unit 28) determines the position of the three-dimensional object and the content image so that the image of the three-dimensional object to which the content image is pasted is substantially a visual image of the real object at the background shooting position. What is necessary is just to set the method to paste. The method of superimposing the content image on the background image is not limited to the above example.
 図11A、図11B及び図11Cは、図10A、図10B及び図10Cに示される背景画像にコンテンツ画像が重畳された合成画像(AR画像)の例をそれぞれ示す図である。
これら図に示されるように、重畳されるコンテンツ画像は、背景画像の撮影位置及び撮影方向に応じて切り替えられる。重畳されるコンテンツ画像における現実オブジェクトの像は、背景画像の撮影位置及び撮影方向から視認される、背景画像が撮影される空間に設置された現実オブジェクトと同等の形態となる。
11A, 11B, and 11C are diagrams illustrating examples of a composite image (AR image) in which a content image is superimposed on the background image illustrated in FIGS. 10A, 10B, and 10C.
As shown in these drawings, the superimposed content image is switched according to the shooting position and shooting direction of the background image. The image of the real object in the superimposed content image has the same form as the real object installed in the space where the background image is captured, which is viewed from the shooting position and shooting direction of the background image.
 図12A、図12B及び図12Cは、AR画像上でのARコンテンツの移動を示す図である。図12Aでは、ARコンテンツがマーカの上に配置されている。図12Bでは、図12Aに示す状態と比較して、ARコンテンツがマーカに対して後ろ方向に移動されている。図12Cでは、図12Aに示す状態と比較して、ARコンテンツがマーカに対して左方向に移動されている。ユーザは、図12Aに示されるAR画像が表示装置15に表示されている状態で、コンテンツ画像を移動する操作を行うことによって、図12B及び図12Cに示されるようなAR画像を見ることができる。また、図12Bでは、コンテンツ画像は、移動後のARコンテンツの位置に応じて、図12Aに示されるコンテンツ画像よりもサイズが縮小されている。図12Cでは、コンテンツ画像は、移動後の位置に応じて、図12Aに示されるコンテンツ画像から切り換えられている。 12A, 12B, and 12C are diagrams illustrating movement of AR content on an AR image. In FIG. 12A, the AR content is arranged on the marker. In FIG. 12B, compared with the state shown in FIG. 12A, the AR content is moved backward with respect to the marker. In FIG. 12C, compared with the state shown in FIG. 12A, the AR content is moved to the left with respect to the marker. The user can view the AR image as shown in FIGS. 12B and 12C by performing an operation of moving the content image in a state where the AR image shown in FIG. 12A is displayed on the display device 15. . In FIG. 12B, the content image is reduced in size compared to the content image shown in FIG. 12A according to the position of the AR content after movement. In FIG. 12C, the content image is switched from the content image shown in FIG. 12A according to the position after movement.
 なお、上述の説明で用いた複数のフローチャートでは、複数の工程(処理)が順番に記載されているが、各実施形態で実行される工程の実行順序は、その記載の順番に制限されない。各実施形態では、図示される工程の順番を内容的に支障のない範囲で変更することができる。また、上述の各実施形態及び各変形例は、内容が相反しない範囲で組み合わせることができる。 In the plurality of flowcharts used in the above description, a plurality of steps (processes) are described in order, but the execution order of the steps executed in each embodiment is not limited to the description order. In each embodiment, the order of the illustrated steps can be changed within a range that does not hinder the contents. Moreover, each above-mentioned embodiment and each modification can be combined in the range with which the content does not conflict.
 上記の各実施形態の一部又は全部は、以下の付記のようにも特定され得る。但し、各実施形態は、以下の記載に限定されない。 Part or all of the above embodiments can be specified as in the following supplementary notes. However, each embodiment is not limited to the following description.
 (付記1)
 画像ファイルの選択操作を受け付ける操作受付部と、
 マーカが写る背景画像を保存する画像ファイルを保持する背景保持部から、前記操作受付部により受け付けられる選択操作で選択された画像ファイルの画像を背景画像として取得する背景取得部と、
 前記背景取得部によって取得される前記背景画像からマーカを検出するマーカ検出部と、
 前記検出されたマーカに関する現実情報及び画像情報に基づいて、3次元座標空間を設定し、かつ、背景撮影位置及び背景撮影方向を算出する基準算出部と、
 表示対象のAR(Augmented Reality)コンテンツを仮想的に配置する前記3次元座標空間上の位置を決定する位置決定部と、
 前記3次元座標空間の前記決定された位置に配置される前記ARコンテンツの前記背景撮影位置及び前記背景撮影方向からの視認画像が前記背景画像に重畳された合成画像を表示部に表示させる表示制御部と、
 を備える画像処理装置。
 (付記2)
 前記ARコンテンツに相当する現実オブジェクトを複数の方向から撮影して得られた複数のコンテンツ画像を各コンテンツ撮影方向に関連付けて保持するコンテンツ保持部から、前記位置決定部によって決定された位置から前記背景撮影位置への方向と近い該コンテンツ撮影方向に関連付けられて保持されるコンテンツ画像を抽出する画像抽出部と、
 前記画像抽出部により抽出されるコンテンツ画像を前記視認画像として前記背景画像に重畳する合成処理部と、
 を更に備える付記1に記載の画像処理装置。
 (付記3)
 前記操作受付部は、前記合成画像に対する前記コンテンツ画像の移動操作を更に受け付け、
 前記位置決定部は、前記受け付けられた移動操作が示す前記3次元座標空間上の前記ARコンテンツの移動後の位置を決定し、
 前記画像抽出部は、前記位置決定部により決定される移動後の位置から前記背景撮影位置への方向と近い前記コンテンツ撮影方向に関連付けられて前記コンテンツ保持部に保持されるコンテンツ画像を新たに抽出し、
 前記合成処理部は、前記背景画像に重畳するコンテンツ画像を前記新たに抽出されたコンテンツ画像に切り換える、
 付記2に記載の画像処理装置。
 (付記4)
 前記現実オブジェクトの現実サイズと、前記マーカ検出部によって検出されたマーカに関する前記現実情報及び前記画像情報と、前記位置決定部により決定される位置とに基づいて、前記抽出されたコンテンツ画像を拡大又は縮小するサイズ調整部、
 を更に備え、
 前記合成処理部は、前記サイズ調整部により調整されたコンテンツ画像を前記視認画像として前記背景画像に重畳する、
 付記2又は3に記載の画像処理装置。
 (付記5)
 前記コンテンツ保持部は、前記各コンテンツ撮影方向に関連付けられる前記複数のコンテンツ画像を複数の現実オブジェクトの各々についてそれぞれ保持し、
 前記操作受付部は、前記複数の現実オブジェクトの中から前記表示対象とする1つの現実オブジェクトを選択する操作を更に受け付け、
 前記画像抽出部は、前記受け付けられた操作により選択された現実オブジェクトに関連して前記コンテンツ保持部に保持される前記複数のコンテンツ画像の中から1つのコンテンツ画像を抽出する、
 付記2から4のいずれか1つに記載の画像処理装置。
 (付記6)
 前記表示制御部は、前記マーカ検出部により前記背景画像から前記マーカが検出されない場合に、マーカが検出されないことを前記表示部に表示させ、
 前記操作受付部は、前記マーカが検出されなかった前記画像ファイル以外の画像ファイルを選択し直す操作を受け付け、
 前記背景取得部は、前記操作により新たに選択された画像ファイルの画像を前記背景画像として新たに取得する、
 付記1から5のいずれか1つに記載の画像処理装置。
 (付記7)
 前記操作受付部は、撮像部からのリアルタイム映像の取得を要求する操作を受け付け、 前記背景取得部は、前記操作受付部によって受け付けられた前記操作に応じて、前記撮像部から得られる前記リアルタイム映像の画像フレームを前記背景画像として逐次取得し、
 前記表示制御部は、前記コンテンツ画像が前記視認画像として重畳されたリアルタイム映像を前記表示部に表示させる、
 付記2から5のいずれか1つに記載の画像処理装置。
 (付記8)
 前記操作受付部は、前記視認画像が重畳されたリアルタイム映像が前記表示部に表示されている状態におけるスナップショット操作を受け付け、
 前記背景取得部は、前記操作受付部による前記スナップショット操作の受け付けに応じて、前記画像フレームを前記背景画像として固定的に保持し、
 前記表示制御部は、前記視認画像が重畳されたリアルタイム映像の表示に替え、前記固定的に保持される背景画像に前記コンテンツ画像が前記視認画像として重畳された前記合成画像を前記表示部に表示させる、
 付記7に記載の画像処理装置。
 (付記9)
 前記マーカ検出部は、形状情報及びサイズ情報をそれぞれ含む複数種のマーカの情報を保持するマーカ情報保持部から、1種のマーカの形状情報及びサイズ情報を取得し、該取得された形状情報及びサイズ情報を用いて、前記マーカを検出する、
 付記1から8のいずれか1つに記載の画像処理装置。
(Appendix 1)
An operation reception unit for receiving an image file selection operation;
A background acquisition unit for acquiring, as a background image, an image of an image file selected by a selection operation received by the operation reception unit, from a background holding unit for storing an image file storing a background image in which a marker is reflected;
A marker detection unit for detecting a marker from the background image acquired by the background acquisition unit;
A reference calculation unit that sets a three-dimensional coordinate space based on real information and image information regarding the detected marker, and calculates a background shooting position and a background shooting direction;
A position determining unit that determines a position in the three-dimensional coordinate space in which AR (Augmented Reality) content to be displayed is virtually arranged;
Display control for causing a display unit to display a composite image in which a visual image from the background photographing position and the background photographing direction of the AR content arranged at the determined position in the three-dimensional coordinate space is superimposed on the background image. And
An image processing apparatus comprising:
(Appendix 2)
From a content holding unit that holds a plurality of content images obtained by shooting a real object corresponding to the AR content from a plurality of directions in association with each content shooting direction, from the position determined by the position determination unit, the background An image extraction unit for extracting a content image held in association with the content shooting direction close to the direction to the shooting position;
A composition processing unit that superimposes the content image extracted by the image extraction unit on the background image as the visual image;
The image processing apparatus according to appendix 1, further comprising:
(Appendix 3)
The operation accepting unit further accepts an operation for moving the content image with respect to the composite image,
The position determination unit determines a position after movement of the AR content in the three-dimensional coordinate space indicated by the accepted movement operation;
The image extraction unit newly extracts a content image held in the content holding unit in association with the content shooting direction close to the direction from the moved position determined by the position determination unit to the background shooting position. And
The composition processing unit switches the content image to be superimposed on the background image to the newly extracted content image.
The image processing apparatus according to appendix 2.
(Appendix 4)
Enlarging or extracting the extracted content image based on the real size of the real object, the real information and the image information regarding the marker detected by the marker detection unit, and the position determined by the position determination unit Size adjustment part to reduce,
Further comprising
The composition processing unit superimposes the content image adjusted by the size adjusting unit on the background image as the visual image.
The image processing apparatus according to appendix 2 or 3.
(Appendix 5)
The content holding unit holds the plurality of content images associated with each content shooting direction for each of a plurality of real objects,
The operation accepting unit further accepts an operation of selecting one real object to be displayed from the plurality of real objects,
The image extraction unit extracts one content image from the plurality of content images held in the content holding unit in relation to a real object selected by the accepted operation;
The image processing apparatus according to any one of appendices 2 to 4.
(Appendix 6)
The display control unit causes the display unit to display that no marker is detected when the marker detection unit does not detect the marker from the background image,
The operation accepting unit accepts an operation of reselecting an image file other than the image file in which the marker is not detected,
The background acquisition unit newly acquires an image of an image file newly selected by the operation as the background image.
The image processing apparatus according to any one of appendices 1 to 5.
(Appendix 7)
The operation accepting unit accepts an operation requesting acquisition of a real-time video from the imaging unit, and the background acquisition unit is the real-time video obtained from the imaging unit in response to the operation accepted by the operation accepting unit. Sequentially obtained as the background image,
The display control unit causes the display unit to display a real-time video in which the content image is superimposed as the visual image.
The image processing apparatus according to any one of appendices 2 to 5.
(Appendix 8)
The operation accepting unit accepts a snapshot operation in a state where a real-time video on which the visual image is superimposed is displayed on the display unit,
The background acquisition unit fixedly holds the image frame as the background image in response to reception of the snapshot operation by the operation reception unit,
The display control unit displays the composite image in which the content image is superimposed as the visual image on the fixedly held background image on the display unit, instead of displaying the real-time video on which the visual image is superimposed. Let
The image processing apparatus according to appendix 7.
(Appendix 9)
The marker detection unit acquires shape information and size information of one type of marker from a marker information holding unit that holds information of a plurality of types of markers each including shape information and size information, and the acquired shape information and Detecting the marker using size information;
The image processing apparatus according to any one of appendices 1 to 8.
 (付記10)
 少なくとも1つのコンピュータにより実行される画像処理方法において、
 画像ファイルの選択操作を受け付け、
 マーカが写る背景画像を保存する画像ファイルを保持する背景保持部から、前記受け付けられた選択操作で選択された画像ファイルの画像を背景画像として取得し、
 前記取得された背景画像からマーカを検出し、
 前記検出されたマーカに関する現実情報及び画像情報に基づいて、3次元座標空間を設定し、
 前記検出されたマーカに関する現実情報及び画像情報に基づいて、背景撮影位置及び背景撮影方向を算出し、
 表示対象のAR(Augmented Reality)コンテンツを仮想的に配置する前記3次元座標空間上の位置を決定し、
 前記3次元座標空間の前記決定された位置に配置される前記ARコンテンツの前記背景撮影位置及び前記背景撮影方向からの視認画像が前記背景画像に重畳された合成画像を表示部に表示させる、
 ことを含む画像処理方法。
 (付記11)
 前記ARコンテンツに相当する現実オブジェクトを複数の方向から撮影して得られた複数のコンテンツ画像を各コンテンツ撮影方向に関連付けて保持するコンテンツ保持部から、前記3次元座標空間上の前記決定された位置から前記背景撮影位置への方向と近い該コンテンツ撮影方向に関連付けられて保持されるコンテンツ画像を抽出し、
 前記抽出されたコンテンツ画像を前記視認画像として前記背景画像に重畳する、
 ことを更に含む付記10に記載の画像処理方法。
 (付記12)
 前記合成画像に対する前記コンテンツ画像の移動操作を受け付け、
 前記受け付けられた移動操作が示す前記3次元座標空間上の前記ARコンテンツの移動後の位置を決定し、
 前記決定された移動後の位置から前記背景撮影位置への方向と近い前記コンテンツ撮影方向に関連付けられて前記コンテンツ保持部に保持されるコンテンツ画像を新たに抽出し、
 前記背景画像に重畳するコンテンツ画像を前記新たに抽出されたコンテンツ画像に切り換える、
 ことを更に含む付記11に記載の画像処理方法。
 (付記13)
 前記現実オブジェクトの現実サイズと、前記検出されたマーカに関する前記現実情報及び前記画像情報と、前記3次元座標空間上の前記決定された位置とに基づいて、前記抽出されたコンテンツ画像を拡大又は縮小する、
 ことを更に含む付記11又は12に記載の画像処理方法。
 (付記14)
 複数の現実オブジェクトの中から前記表示対象とする1つの現実オブジェクトを選択する操作を受け付ける、
 ことを更に含み、
 前記コンテンツ保持部は、前記各コンテンツ撮影方向に関連付けられる前記複数のコンテンツ画像を前記複数の現実オブジェクトの各々についてそれぞれ保持し、
 前記コンテンツ画像の抽出は、前記受け付けられた操作により選択された現実オブジェクトに関連して前記コンテンツ保持部に保持される前記複数のコンテンツ画像の中から1つのコンテンツ画像を抽出する、
 付記11から13のいずれか1つに記載の画像処理方法。
 (付記15)
 前記背景画像から前記マーカが検出されない場合に、マーカが検出されないことを前記表示部に表示させ、
 前記マーカが検出されなかった前記画像ファイル以外の画像ファイルを選択し直す操作を受け付け、
 前記操作により新たに選択された画像ファイルの画像を前記背景画像として新たに取得する、
 ことを更に含む付記10から14のいずれか1つに記載の画像処理方法。
 (付記16)
 撮像部からのリアルタイム映像の取得を要求する操作を受け付け、
 前記受け付けられた前記操作に応じて、前記撮像部から得られる前記リアルタイム映像の画像フレームを前記背景画像として逐次取得し、
 前記コンテンツ画像が前記視認画像として重畳されたリアルタイム映像を前記表示部に表示させる、
 ことを更に含む付記11から14のいずれか1つに記載の画像処理方法。
 (付記17)
 前記視認画像が重畳されたリアルタイム映像が前記表示部に表示されている状態におけるスナップショット操作を受け付け、
 前記スナップショット操作の受け付けに応じて、前記画像フレームを前記背景画像として固定的に保持し、
 前記視認画像が重畳されたリアルタイム映像の表示に替え、前記固定的に保持される背景画像に前記コンテンツ画像が前記視認画像として重畳された前記合成画像を前記表示部に表示させる、
 ことを更に含む付記16に記載の画像処理方法。
 (付記18)
 形状情報及びサイズ情報をそれぞれ含む複数種のマーカの情報を保持するマーカ情報保持部から、1種のマーカの形状情報及びサイズ情報を取得する、
 ことを更に含み、
 前記マーカの検出は、前記取得された形状情報及びサイズ情報を用いて、前記マーカを検出する、
 付記10から17のいずれか1つに記載の画像処理方法。
(Appendix 10)
In an image processing method executed by at least one computer,
Accept image file selection operations,
An image of the image file selected by the accepted selection operation is acquired as a background image from a background holding unit that holds an image file that stores a background image showing a marker,
Detecting a marker from the acquired background image;
Based on real information and image information regarding the detected marker, a three-dimensional coordinate space is set,
Based on the actual information and image information about the detected marker, calculate a background shooting position and a background shooting direction,
Determining a position in the three-dimensional coordinate space where the AR (Augmented Reality) content to be displayed is virtually arranged;
Displaying a composite image in which a visual image from the background photographing position and the background photographing direction of the AR content arranged at the determined position in the three-dimensional coordinate space is superimposed on the background image on a display unit;
An image processing method.
(Appendix 11)
The determined position in the three-dimensional coordinate space from a content holding unit that holds a plurality of content images obtained by shooting a real object corresponding to the AR content from a plurality of directions in association with each content shooting direction. A content image held in association with the content shooting direction that is close to the direction from the background shooting position to
Superimposing the extracted content image on the background image as the visual image;
The image processing method according to appendix 10, further including:
(Appendix 12)
Accepting an operation of moving the content image with respect to the composite image;
Determining a position after movement of the AR content in the three-dimensional coordinate space indicated by the accepted movement operation;
Newly extracting a content image held in the content holding unit in association with the content shooting direction close to the direction from the determined moved position to the background shooting position,
Switching the content image to be superimposed on the background image to the newly extracted content image;
The image processing method according to claim 11, further comprising:
(Appendix 13)
The extracted content image is enlarged or reduced based on the real size of the real object, the real information and the image information regarding the detected marker, and the determined position in the three-dimensional coordinate space. To
The image processing method according to appendix 11 or 12, further including:
(Appendix 14)
Receiving an operation of selecting one real object to be displayed from among a plurality of real objects;
Further including
The content holding unit holds the plurality of content images associated with the content shooting directions for each of the plurality of real objects,
In the extraction of the content image, one content image is extracted from the plurality of content images held in the content holding unit in relation to the real object selected by the accepted operation.
The image processing method according to any one of appendices 11 to 13.
(Appendix 15)
If the marker is not detected from the background image, display that the marker is not detected on the display unit,
Accepting an operation to reselect an image file other than the image file in which the marker is not detected,
An image of the image file newly selected by the operation is newly acquired as the background image.
The image processing method according to any one of appendices 10 to 14, further including:
(Appendix 16)
Accepting an operation requesting acquisition of real-time video from the imaging unit,
In response to the accepted operation, the image frame of the real-time video obtained from the imaging unit is sequentially acquired as the background image,
Displaying the real-time video on which the content image is superimposed as the visual image on the display unit;
The image processing method according to any one of appendices 11 to 14, further including:
(Appendix 17)
Accepting a snapshot operation in a state where the real-time video on which the visual image is superimposed is displayed on the display unit,
In response to accepting the snapshot operation, the image frame is fixedly held as the background image,
Instead of displaying the real-time video on which the visual image is superimposed, the composite image in which the content image is superimposed as the visual image on the fixedly held background image is displayed on the display unit.
The image processing method according to appendix 16, further including:
(Appendix 18)
Obtaining shape information and size information of one type of marker from a marker information holding unit that holds information of a plurality of types of markers each including shape information and size information;
Further including
The detection of the marker detects the marker using the acquired shape information and size information.
18. The image processing method according to any one of appendices 10 to 17.
 (付記19)
 撮像部から得られるリアルタイム映像の画像フレームを背景画像として逐次取得する背景取得部と、
 前記背景取得部によって取得される前記背景画像からマーカを検出するマーカ検出部と、
 前記検出されたマーカに関する現実情報及び画像情報に基づいて、3次元座標空間を設定し、かつ、背景撮影位置及び背景撮影方向を算出する基準算出部と、
 表示対象のAR(Augmented Reality)コンテンツを仮想的に配置する前記3次元座標空間上の位置を決定する位置決定部と、
 前記ARコンテンツに相当する現実オブジェクトを複数の方向から撮影して得られた複数のコンテンツ画像を各コンテンツ撮影方向に関連付けて保持するコンテンツ保持部から、前記位置決定部によって決定された位置から前記背景撮影位置への方向と近い該コンテンツ撮影方向に関連付けられて保持されるコンテンツ画像を抽出する画像抽出部と、
 前記画像抽出部により抽出されるコンテンツ画像を前記視認画像として前記背景画像に重畳する合成処理部と、
 前記コンテンツ画像が前記視認画像として重畳されたリアルタイム映像を表示部に表示させる表示制御部と、
 を備える画像処理装置。
 (付記20)
 前記視認画像が重畳されたリアルタイム映像が前記表示部に表示されている状態におけるスナップショット操作を受け付ける操作受付部、
 を更に備え、
 前記背景取得部は、前記操作受付部による前記スナップショット操作の受け付けに応じて、前記画像フレームを前記背景画像として固定的に保持し、
 前記表示制御部は、前記視認画像が重畳されたリアルタイム映像の表示に替え、前記固定的に保持される背景画像に前記コンテンツ画像が前記視認画像として重畳された前記合成画像を前記表示部に表示させる、
 付記19に記載の画像処理装置。
(Appendix 19)
A background acquisition unit that sequentially acquires an image frame of a real-time video obtained from the imaging unit as a background image;
A marker detection unit for detecting a marker from the background image acquired by the background acquisition unit;
A reference calculation unit that sets a three-dimensional coordinate space based on real information and image information regarding the detected marker, and calculates a background shooting position and a background shooting direction;
A position determining unit that determines a position in the three-dimensional coordinate space in which AR (Augmented Reality) content to be displayed is virtually arranged;
From a content holding unit that holds a plurality of content images obtained by shooting a real object corresponding to the AR content from a plurality of directions in association with each content shooting direction, from the position determined by the position determination unit, the background An image extraction unit for extracting a content image held in association with the content shooting direction close to the direction to the shooting position;
A composition processing unit that superimposes the content image extracted by the image extraction unit on the background image as the visual image;
A display control unit that causes a display unit to display a real-time video in which the content image is superimposed as the visual image;
An image processing apparatus comprising:
(Appendix 20)
An operation accepting unit that accepts a snapshot operation in a state where the real-time video on which the visual image is superimposed is displayed on the display unit;
Further comprising
The background acquisition unit fixedly holds the image frame as the background image in response to reception of the snapshot operation by the operation reception unit,
The display control unit displays the composite image in which the content image is superimposed as the visual image on the fixedly held background image on the display unit, instead of displaying the real-time video on which the visual image is superimposed. Let
The image processing apparatus according to appendix 19.
 (付記21)
 撮像部から得られるリアルタイム映像の画像フレームを背景画像として逐次取得し、
 前記取得された背景画像からマーカを検出し、
 前記検出されたマーカに関する現実情報及び画像情報に基づいて、3次元座標空間を設定し、
 前記検出されたマーカに関する現実情報及び画像情報に基づいて、前記背景画像の背景撮影位置及び背景撮影方向を算出し、
 表示対象のAR(Augmented Reality)コンテンツを仮想的に配置する前記3次元座標空間上の位置を決定し、
 前記ARコンテンツに相当する現実オブジェクトを複数の方向から撮影して得られた複数のコンテンツ画像を各コンテンツ撮影方向に関連付けて保持するコンテンツ保持部から、前記3次元座標空間上の前記決定された位置から前記背景撮影位置への方向と近い該コンテンツ撮影方向に関連付けられて保持されるコンテンツ画像を抽出し、
 前記抽出されたコンテンツ画像を前記視認画像として前記背景画像に重畳し、
 前記コンテンツ画像が前記視認画像として重畳されたリアルタイム映像を表示部に表示させる、
 ことを含む画像処理方法。
 (付記22)
 前記視認画像が重畳されたリアルタイム映像が前記表示部に表示されている状態におけるスナップショット操作を受け付け、
 前記スナップショット操作の受け付けに応じて、前記画像フレームを前記背景画像として固定的に保持し、
 前記視認画像が重畳されたリアルタイム映像の表示に替え、前記固定的に保持される背景画像に前記コンテンツ画像が前記視認画像として重畳された前記合成画像を前記表示部に表示させる、
 ことを更に含む付記21に記載の画像処理方法。
(Appendix 21)
Obtain sequentially the image frame of the real-time video obtained from the imaging unit as a background image,
Detecting a marker from the acquired background image;
Based on real information and image information regarding the detected marker, a three-dimensional coordinate space is set,
Based on the actual information and image information regarding the detected marker, calculate a background shooting position and a background shooting direction of the background image,
Determining a position in the three-dimensional coordinate space where the AR (Augmented Reality) content to be displayed is virtually arranged;
The determined position in the three-dimensional coordinate space from a content holding unit that holds a plurality of content images obtained by shooting a real object corresponding to the AR content from a plurality of directions in association with each content shooting direction. A content image held in association with the content shooting direction that is close to the direction from the background shooting position to
Superimposing the extracted content image as the visual image on the background image;
Displaying the real-time video on which the content image is superimposed as the visual image on the display unit;
An image processing method.
(Appendix 22)
Accepting a snapshot operation in a state where the real-time video on which the visual image is superimposed is displayed on the display unit,
In response to accepting the snapshot operation, the image frame is fixedly held as the background image,
Instead of displaying the real-time video on which the visual image is superimposed, the composite image in which the content image is superimposed as the visual image on the fixedly held background image is displayed on the display unit.
The image processing method according to claim 21, further comprising:
 (付記23)
 付記10から18、21及び22のいずれか1つに記載の画像処理方法を少なくとも1つのコンピュータに実行させるプログラム。
 以上、実施形態を参照して本願発明を説明したが、本願発明は上記実施形態に限定されるものではない。本願発明の構成や詳細には、本願発明のスコープ内で当業者が理解し得る様々な変更をすることができる。
 この出願は、2013年11月12日に出願された日本出願特願2013-233805を基礎とする優先権を主張し、その開示の全てをここに取り込む。
(Appendix 23)
A program that causes at least one computer to execute the image processing method according to any one of Supplementary Notes 10 to 18, 21, and 22.
While the present invention has been described with reference to the embodiments, the present invention is not limited to the above embodiments. Various changes that can be understood by those skilled in the art can be made to the configuration and details of the present invention within the scope of the present invention.
This application claims the priority on the basis of Japanese application Japanese Patent Application No. 2013-233805 for which it applied on November 12, 2013, and takes in those the indications of all here.
 10、10A、100  画像処理装置
 11  CPU
 12  メモリ
 15  表示装置
 16  入力装置
 17  カメラ
 18  可搬型記録媒体
 21、101  操作受付部
 22、102  背景取得部
 23、103  マーカ検出部
 24、104  基準算出部
 25、105  位置決定部
 26  画像抽出部
 27  サイズ調整部
 28  合成処理部
 29、106  表示制御部
 31、107  背景保持部
 32  コンテンツ保持部
 35  マーカ情報保持部
 108  表示部
10, 10A, 100 Image processing apparatus 11 CPU
DESCRIPTION OF SYMBOLS 12 Memory 15 Display apparatus 16 Input apparatus 17 Camera 18 Portable recording medium 21, 101 Operation reception part 22, 102 Background acquisition part 23, 103 Marker detection part 24, 104 Reference | standard calculation part 25, 105 Position determination part 26 Image extraction part 27 Size adjustment unit 28 Composition processing unit 29, 106 Display control unit 31, 107 Background holding unit 32 Content holding unit 35 Marker information holding unit 108 Display unit

Claims (11)

  1.  画像ファイルの選択操作を受け付ける操作受付手段と、
     マーカが写る背景画像を保存する画像ファイルを保持する背景保持手段から、前記操作受付手段により受け付けられる選択操作で選択された画像ファイルの画像を背景画像として取得する背景取得手段と、
     前記背景取得手段によって取得される前記背景画像からマーカを検出するマーカ検出手段と、
     前記検出されたマーカに関する現実情報及び画像情報に基づいて、3次元座標空間を設定し、かつ、背景撮影位置及び背景撮影方向を算出する基準算出手段と、
     表示対象のAR(Augmented Reality)コンテンツを仮想的に配置する前記3次元座標空間上の位置を決定する位置決定手段と、
     前記3次元座標空間の前記決定された位置に配置される前記ARコンテンツの前記背景撮影位置及び前記背景撮影方向からの視認画像が前記背景画像に重畳された合成画像を表示手段に表示させる表示制御手段と、
     を備える画像処理装置。
    Operation accepting means for accepting an image file selection operation;
    A background acquisition unit that acquires an image of an image file selected by a selection operation received by the operation reception unit as a background image from a background holding unit that stores an image file that stores a background image showing a marker;
    Marker detection means for detecting a marker from the background image acquired by the background acquisition means;
    Reference calculation means for setting a three-dimensional coordinate space and calculating a background shooting position and a background shooting direction based on the actual information and image information regarding the detected marker;
    Position determining means for determining a position in the three-dimensional coordinate space in which AR (Augmented Reality) content to be displayed is virtually arranged;
    Display control for causing a display unit to display a composite image in which a visual image from the background photographing position and the background photographing direction of the AR content arranged at the determined position in the three-dimensional coordinate space is superimposed on the background image. Means,
    An image processing apparatus comprising:
  2.  前記ARコンテンツに相当する現実オブジェクトを複数の方向から撮影して得られた複数のコンテンツ画像を各コンテンツ撮影方向に関連付けて保持するコンテンツ保持手段から、前記位置決定手段で決定された位置から前記背景撮影位置への方向と近い該コンテンツ撮影方向に関連付けられて保持されるコンテンツ画像を抽出する画像抽出手段と、
     前記画像抽出手段により抽出されるコンテンツ画像を前記視認画像として前記背景画像に重畳する合成処理手段と、
     を更に備える請求項1に記載の画像処理装置。
    From a content holding unit that holds a plurality of content images obtained by shooting a real object corresponding to the AR content from a plurality of directions in association with each content shooting direction, from the position determined by the position determination unit, the background Image extracting means for extracting a content image held in association with the content shooting direction close to the direction to the shooting position;
    Composition processing means for superimposing the content image extracted by the image extraction means on the background image as the visual image;
    The image processing apparatus according to claim 1, further comprising:
  3.  前記操作受付手段は、前記合成画像に対する前記コンテンツ画像の移動操作を更に受け付け、
     前記位置決定手段は、前記受け付けられた移動操作が示す前記3次元座標空間上の前記ARコンテンツの移動後の位置を決定し、
     前記画像抽出手段は、前記位置決定手段により決定される移動後の位置から前記背景撮影位置への方向と近い前記コンテンツ撮影方向に関連付けられて前記コンテンツ保持手段に保持されるコンテンツ画像を新たに抽出し、
     前記合成処理手段は、前記背景画像に重畳するコンテンツ画像を前記新たに抽出されたコンテンツ画像に切り換える、
     請求項2に記載の画像処理装置。
    The operation accepting unit further accepts a moving operation of the content image with respect to the composite image,
    The position determining means determines a position after movement of the AR content on the three-dimensional coordinate space indicated by the accepted movement operation;
    The image extraction unit newly extracts a content image held in the content holding unit in association with the content shooting direction close to the direction from the moved position determined by the position determination unit to the background shooting position. And
    The composition processing means switches the content image to be superimposed on the background image to the newly extracted content image.
    The image processing apparatus according to claim 2.
  4.  前記現実オブジェクトの現実サイズと、前記マーカ検出手段によって検出されたマーカに関する前記現実情報及び前記画像情報と、前記位置決定手段により決定される位置とに基づいて、前記抽出されたコンテンツ画像を拡大又は縮小するサイズ調整手段、
     を更に備え、
     前記合成処理手段は、前記サイズ調整手段により調整されたコンテンツ画像を前記視認画像として前記背景画像に重畳する、
     請求項2又は3に記載の画像処理装置。
    Enlarging or extracting the extracted content image based on the real size of the real object, the real information and the image information relating to the marker detected by the marker detection means, and the position determined by the position determination means Size adjustment means to reduce,
    Further comprising
    The composition processing unit superimposes the content image adjusted by the size adjusting unit on the background image as the visual image;
    The image processing apparatus according to claim 2.
  5.  前記コンテンツ保持手段は、前記各コンテンツ撮影方向に関連付けられる前記複数のコンテンツ画像を複数の現実オブジェクトの各々についてそれぞれ保持し、
     前記操作受付手段は、前記複数の現実オブジェクトの中から前記表示対象とする1つの現実オブジェクトを選択する操作を更に受け付け、
     前記画像抽出手段は、前記受け付けられた操作により選択された現実オブジェクトに関連して前記コンテンツ保持手段に保持される前記複数のコンテンツ画像の中から1つのコンテンツ画像を抽出する、
     請求項2から4のいずれか1項に記載の画像処理装置。
    The content holding unit holds the plurality of content images associated with each content shooting direction for each of a plurality of real objects,
    The operation accepting unit further accepts an operation of selecting one real object to be displayed from the plurality of real objects,
    The image extracting means extracts one content image from the plurality of content images held in the content holding means in relation to a real object selected by the accepted operation;
    The image processing apparatus according to claim 2.
  6.  前記表示制御手段は、前記マーカ検出手段により前記背景画像から前記マーカが検出されない場合に、マーカが検出されないことを前記表示手段に表示させ、
     前記操作受付手段は、前記マーカが検出されなかった前記画像ファイル以外の画像ファイルを選択し直す操作を受け付け、
     前記背景取得手段は、前記操作により新たに選択された画像ファイルの画像を前記背景画像として新たに取得する、
     請求項1から5のいずれか1項に記載の画像処理装置。
    The display control means causes the display means to display that no marker is detected when the marker is not detected from the background image by the marker detection means,
    The operation accepting unit accepts an operation of reselecting an image file other than the image file in which the marker is not detected,
    The background acquisition unit newly acquires an image of an image file newly selected by the operation as the background image.
    The image processing apparatus according to claim 1.
  7.  前記操作受付手段は、撮像手段からのリアルタイム映像の取得を要求する操作を受け付け、
     前記背景取得手段は、前記操作受付手段で受け付けられた前記操作に応じて、前記撮像手段から得られる前記リアルタイム映像の画像フレームを前記背景画像として逐次取得し、
     前記表示制御手段は、前記コンテンツ画像が前記視認画像として重畳されたリアルタイム映像を前記表示手段に表示させる、
     請求項2から5のいずれか1項に記載の画像処理装置。
    The operation accepting means accepts an operation requesting acquisition of real-time video from the imaging means,
    The background acquisition unit sequentially acquires an image frame of the real-time video obtained from the imaging unit as the background image in response to the operation received by the operation reception unit,
    The display control means causes the display means to display a real-time video in which the content image is superimposed as the visual image.
    The image processing apparatus according to claim 2.
  8.  前記操作受付手段は、前記視認画像が重畳されたリアルタイム映像が前記表示手段に表示されている状態におけるスナップショット操作を受け付け、
     前記背景取得手段は、前記操作受付手段による前記スナップショット操作の受け付けに応じて、前記画像フレームを前記背景画像として固定的に保持し、
     前記表示制御手段は、前記視認画像が重畳されたリアルタイム映像の表示に替え、前記固定的に保持される背景画像に前記コンテンツ画像が前記視認画像として重畳された前記合成画像を前記表示手段に表示させる、
     請求項7に記載の画像処理装置。
    The operation accepting unit accepts a snapshot operation in a state where a real-time video on which the visual image is superimposed is displayed on the display unit,
    The background acquisition unit holds the image frame fixedly as the background image in response to reception of the snapshot operation by the operation reception unit,
    The display control unit displays the composite image in which the content image is superimposed as the visual image on the fixedly held background image on the display unit, instead of displaying the real-time video on which the visual image is superimposed. Let
    The image processing apparatus according to claim 7.
  9.  前記マーカ検出手段は、形状情報及びサイズ情報をそれぞれ含む複数種のマーカの情報を保持するマーカ情報保持手段から、1種のマーカの形状情報及びサイズ情報を取得し、該取得された形状情報及びサイズ情報を用いて、前記マーカを検出する、
     請求項1から8のいずれか1項に記載の画像処理装置。
    The marker detection means acquires shape information and size information of one type of marker from marker information holding means for holding information of a plurality of types of markers each including shape information and size information, and the acquired shape information and Detecting the marker using size information;
    The image processing apparatus according to claim 1.
  10.  画像ファイルの選択操作を受け付け、
     マーカが写る背景画像を保存する画像ファイルを保持する背景保持手段から、前記受け付けられた選択操作で選択された画像ファイルの画像を背景画像として取得し、
     前記取得された背景画像からマーカを検出し、
     前記検出されたマーカに関する現実情報及び画像情報に基づいて、3次元座標空間を設定し、
     前記検出されたマーカに関する現実情報及び画像情報に基づいて、背景撮影位置及び背景撮影方向を算出し、
     表示対象のAR(Augmented Reality)コンテンツを仮想的に配置する前記3次元座標空間上の位置を決定し、
     前記3次元座標空間の前記決定された位置に配置される前記ARコンテンツの前記背景撮影位置及び前記背景撮影方向からの視認画像が前記背景画像に重畳された合成画像を表示手段に表示させる、
     画像処理方法。
    Accept image file selection operations,
    From a background holding unit that holds an image file that stores a background image showing a marker, an image of the image file selected by the accepted selection operation is acquired as a background image,
    Detecting a marker from the acquired background image;
    Based on real information and image information regarding the detected marker, a three-dimensional coordinate space is set,
    Based on the actual information and image information about the detected marker, calculate a background shooting position and a background shooting direction,
    Determining a position in the three-dimensional coordinate space where the AR (Augmented Reality) content to be displayed is virtually arranged;
    Displaying a composite image in which a visual image from the background photographing position and the background photographing direction of the AR content arranged at the determined position in the three-dimensional coordinate space is superimposed on the background image on a display unit;
    Image processing method.
  11.  コンピュータを、
     画像ファイルの選択操作を受け付ける操作受付手段と、
     マーカが写る背景画像を保存する画像ファイルを保持する背景保持部から、前記操作受付手段により受け付けられる選択操作で選択された画像ファイルの画像を背景画像として取得する背景取得手段と、
     前記背景取得手段によって取得される前記背景画像からマーカを検出するマーカ検出手段と、
     前記検出されたマーカに関する現実情報及び画像情報に基づいて、3次元座標空間を設定し、かつ、背景撮影位置及び背景撮影方向を算出する基準算出手段と、
     表示対象のAR(Augmented Reality)コンテンツを仮想的に配置する前記3次元座標空間上の位置を決定する位置決定手段と、
     前記3次元座標空間の前記決定された位置に配置される前記ARコンテンツの前記背景撮影位置及び前記背景撮影方向からの視認画像が前記背景画像に重畳された合成画像を表示手段に表示させる表示制御手段と、
     して動作させるプログラムを記憶する記録媒体。
    Computer
    Operation accepting means for accepting an image file selection operation;
    A background acquisition unit that acquires, as a background image, an image of an image file selected by a selection operation received by the operation reception unit from a background holding unit that stores an image file that stores a background image in which a marker is reflected;
    Marker detection means for detecting a marker from the background image acquired by the background acquisition means;
    Reference calculation means for setting a three-dimensional coordinate space and calculating a background shooting position and a background shooting direction based on the actual information and image information regarding the detected marker;
    Position determining means for determining a position in the three-dimensional coordinate space in which AR (Augmented Reality) content to be displayed is virtually arranged;
    Display control for causing a display unit to display a composite image in which a visual image from the background photographing position and the background photographing direction of the AR content arranged at the determined position in the three-dimensional coordinate space is superimposed on the background image. Means,
    Recording medium for storing a program to be operated.
PCT/JP2014/005387 2013-11-12 2014-10-23 Image processing device, image processing method, and program storage medium WO2015072091A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2013-233805 2013-11-12
JP2013233805A JP2017016166A (en) 2013-11-12 2013-11-12 Image processing apparatus and image processing method

Publications (1)

Publication Number Publication Date
WO2015072091A1 true WO2015072091A1 (en) 2015-05-21

Family

ID=53057044

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2014/005387 WO2015072091A1 (en) 2013-11-12 2014-10-23 Image processing device, image processing method, and program storage medium

Country Status (2)

Country Link
JP (1) JP2017016166A (en)
WO (1) WO2015072091A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112672057A (en) * 2020-12-25 2021-04-16 维沃移动通信有限公司 Shooting method and device
US11954804B2 (en) 2018-08-28 2024-04-09 Sony Corporation Information processing device and information processing method to display a virtual object

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2020095645A (en) * 2018-12-14 2020-06-18 ヤフー株式会社 Display program, display device, and display method
JP2020095602A (en) * 2018-12-14 2020-06-18 ヤフー株式会社 Display program, display device, and display method

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005293141A (en) * 2004-03-31 2005-10-20 Canon Inc Method and device for detecting marker, and method for estimating position and attitude
JP2006053694A (en) * 2004-08-10 2006-02-23 Riyuukoku Univ Space simulator, space simulation method, space simulation program and recording medium
JP2012084155A (en) * 2010-10-13 2012-04-26 Toshiba Corp Image processing system, image processing method and image processing program
JP2012108711A (en) * 2010-11-17 2012-06-07 Nec Casio Mobile Communications Ltd Terminal device, operation method of terminal device, and expanded reality system
JP2013183333A (en) * 2012-03-02 2013-09-12 Alpine Electronics Inc Augmented reality system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005293141A (en) * 2004-03-31 2005-10-20 Canon Inc Method and device for detecting marker, and method for estimating position and attitude
JP2006053694A (en) * 2004-08-10 2006-02-23 Riyuukoku Univ Space simulator, space simulation method, space simulation program and recording medium
JP2012084155A (en) * 2010-10-13 2012-04-26 Toshiba Corp Image processing system, image processing method and image processing program
JP2012108711A (en) * 2010-11-17 2012-06-07 Nec Casio Mobile Communications Ltd Terminal device, operation method of terminal device, and expanded reality system
JP2013183333A (en) * 2012-03-02 2013-09-12 Alpine Electronics Inc Augmented reality system

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11954804B2 (en) 2018-08-28 2024-04-09 Sony Corporation Information processing device and information processing method to display a virtual object
CN112672057A (en) * 2020-12-25 2021-04-16 维沃移动通信有限公司 Shooting method and device
CN112672057B (en) * 2020-12-25 2022-07-15 维沃移动通信有限公司 Shooting method and device

Also Published As

Publication number Publication date
JP2017016166A (en) 2017-01-19

Similar Documents

Publication Publication Date Title
US10460512B2 (en) 3D skeletonization using truncated epipolar lines
US9495068B2 (en) Three-dimensional user interface apparatus and three-dimensional operation method
US9541997B2 (en) Three-dimensional user interface apparatus and three-dimensional operation method
US9495802B2 (en) Position identification method and system
US10410089B2 (en) Training assistance using synthetic images
EP2956843B1 (en) Human-body-gesture-based region and volume selection for hmd
JP6167703B2 (en) Display control device, program, and recording medium
US9361731B2 (en) Method and apparatus for displaying video on 3D map
US9268410B2 (en) Image processing device, image processing method, and program
JP2012212345A (en) Terminal device, object control method and program
JP7026825B2 (en) Image processing methods and devices, electronic devices and storage media
US10950056B2 (en) Apparatus and method for generating point cloud data
JP7353782B2 (en) Information processing device, information processing method, and program
KR102450236B1 (en) Electronic apparatus, method for controlling thereof and the computer readable recording medium
US10748000B2 (en) Method, electronic device, and recording medium for notifying of surrounding situation information
WO2015072091A1 (en) Image processing device, image processing method, and program storage medium
JP6290020B2 (en) Image processing apparatus, image processing method, and program
JP6371547B2 (en) Image processing apparatus, method, and program
JP2015184986A (en) Compound sense of reality sharing device
JP6405539B2 (en) Label information processing apparatus for multi-viewpoint image and label information processing method
JP6357412B2 (en) Information processing apparatus, information processing system, information processing method, and program
KR20210076615A (en) Method and appratus for providong video infromation
KR101741149B1 (en) Method and device for controlling a virtual camera&#39;s orientation
JP2019040357A (en) Image processing system, image processing method and computer program
WO2023160072A1 (en) Human-computer interaction method and apparatus in augmented reality (ar) scene, and electronic device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14861688

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14861688

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP