WO2018014601A1 - Method and relevant apparatus for orientational tracking, method and device for realizing augmented reality - Google Patents

Method and relevant apparatus for orientational tracking, method and device for realizing augmented reality Download PDF

Info

Publication number
WO2018014601A1
WO2018014601A1 PCT/CN2017/080276 CN2017080276W WO2018014601A1 WO 2018014601 A1 WO2018014601 A1 WO 2018014601A1 CN 2017080276 W CN2017080276 W CN 2017080276W WO 2018014601 A1 WO2018014601 A1 WO 2018014601A1
Authority
WO
WIPO (PCT)
Prior art keywords
target entity
video frame
visible surface
globe
direction vector
Prior art date
Application number
PCT/CN2017/080276
Other languages
French (fr)
Chinese (zh)
Inventor
刘钢
熊剑明
陈健
方堃
韦晓宁
Original Assignee
央数文化(上海)股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 央数文化(上海)股份有限公司 filed Critical 央数文化(上海)股份有限公司
Publication of WO2018014601A1 publication Critical patent/WO2018014601A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B27/00Planetaria; Globes
    • G09B27/08Globes

Definitions

  • the present application relates to the field of computers, and in particular, to an azimuth tracking method, a method for implementing augmented reality, and related devices and devices.
  • the globe In the process of people knowing the world, the globe has always been an indispensable geographic teaching aid. It allows people to understand the concept of space and location in a more vivid and interesting way, and to understand the vastness of the land, the vastness of the ocean, and the change of the four seasons. Especially for children to form a correct scientific and worldview is also of great significance.
  • traditional globes can only express very limited geographic information, such as the boundary between land and ocean, the border between countries, etc., and for richer geographic information, such as the internal structure of the Earth, the Earth in the solar system. The position of the earth, the distribution of animals and plants on the surface of the earth, etc., the traditional globe is powerless.
  • the purpose of the application is to provide an orientation tracking method, a method for implementing augmented reality, and related devices and devices.
  • the present application provides a method for tracking an orientation of a target entity, the method comprising:
  • the feature set includes a plurality of surface regions of the target entity and image feature information of each surface region;
  • a center point coordinate and a direction vector of the visible surface of the target entity are respectively determined according to a center point coordinate and a direction vector of the surface area matching the visible surface.
  • the method further includes:
  • the target entity is preprocessed to construct a prefabricated feature set of the target entity, including:
  • determining a center point coordinate of the visible surface of the target entity according to a center point coordinate of the surface area matching the visible surface comprises:
  • the center point coordinates of the plurality of surface regions are subjected to weighted averaging processing, and the center point coordinates of the target entity are determined.
  • determining the target entity visible surface direction vector according to a direction vector of the surface area matching the visible surface comprises:
  • the direction vectors of the plurality of surface regions are weighted and added and subjected to vector normalization processing to obtain a direction vector of the visible surface of the target entity.
  • the application also provides a method for implementing augmented reality, the method comprising:
  • the display portion is rendered into the video frame based on the center point coordinates such that a visible surface of the target entity is covered by the display portion.
  • the method further includes:
  • the virtual model is provided with a trigger point
  • the method further includes:
  • the trigger effect of the trigger point is rendered into the video frame.
  • the target entity is a physical globe
  • the virtual model is a virtual globe matching the physical globe.
  • an orientation tracking device for a target entity comprising:
  • An image acquisition module configured to acquire a video frame that includes a target entity
  • a feature extraction module configured to perform image feature extraction on a target entity in the video frame, and acquire image feature information of a visible surface of the target entity
  • a feature matching module configured to perform feature recognition on the acquired image feature information of the visible surface of the target entity based on the set of prefabricated features of the target entity, and acquire a surface region matching the visible surface in the prefabricated feature set
  • the set of prefabricated features includes a plurality of surface regions of the target entity and image feature information of each surface region;
  • an integrated processing module configured to respectively determine a center point coordinate and a direction vector of the visible surface of the target entity according to a center point coordinate and a direction vector of the surface area matching the visible surface.
  • the device further includes:
  • a pre-processing module configured to pre-process the target entity before acquiring the video frame that includes the target entity, and construct a pre-made feature set of the target entity.
  • the preprocessing module is configured to control the image acquisition module to acquire a target real An image of a plurality of surface regions, wherein the plurality of surface regions cover at least all surface regions of the target entity; controlling the feature extraction module to perform image feature extraction on images of the plurality of surface regions to obtain the Image feature information for each surface region; and constructing a set of prefabricated features of the target entity based on image feature information for each of the surface regions.
  • the integrated processing module is configured to perform weighted averaging processing on center point coordinates of the plurality of surface regions when there are multiple surface regions matching the visible surface, and determine a center point coordinate of the target entity .
  • the integrated processing module is configured to weight-add the direction vectors of the plurality of surface regions and perform vector normalization processing to obtain the target when there are multiple surface regions matching the visible surface The direction vector of the solid visible surface.
  • the application also provides a device for realizing augmented reality, the device comprising:
  • An orientation tracking device configured to acquire a center point coordinate and a direction vector of the target entity
  • a rendering device configured to acquire a display portion corresponding to the visible surface from the virtual model according to the direction vector, and render the display portion into the video frame according to the center point coordinate, so that the The visible surface of the target entity is covered by the display portion;
  • An output device for outputting a finished video frame.
  • the rendering device is further configured to acquire an information superposition point in the display portion, and render virtual information corresponding to the information superimposition point into the video frame to make a corresponding position of the display portion The virtual information is displayed.
  • the virtual model is provided with a trigger point
  • the rendering device is further configured to determine whether a trigger point of the display portion is located in a trigger region of the video frame, and when the trigger point is in a trigger region of the video frame, triggering the trigger point Rendered into the video frame.
  • the target entity is a physical globe
  • the virtual model is a virtual globe matching the physical globe.
  • the solution of the present application can superimpose some virtual graphics, images, information, etc. into a real scene based on augmented reality technology, and enhances the real scene by using the physical globe as a target entity. It enables it to combine the advantages of both physical globes and virtual globes in the prior art.
  • Get the central coordinate point of the target entity in real time and The direction vector realizes the azimuth tracking, and the real-time rendering of the output video frame based on the central coordinate point and the direction vector, so that the virtual globe rotates at the same angle and speed during the manipulation of the physical globe, which can greatly improve the user.
  • the direction vector realizes the azimuth tracking, and the real-time rendering of the output video frame based on the central coordinate point and the direction vector, so that the virtual globe rotates at the same angle and speed during the manipulation of the physical globe, which can greatly improve the user.
  • the virtual globe rotates at the same angle and speed during the manipulation of the physical globe, which can greatly improve the user.
  • the image feature extraction is performed by the video frame containing the target entity acquired in real time, and the image feature matching is performed with the pre-built pre-made feature set, and the calculation is performed based on the matching result to obtain the center of the visible surface of the target entity in the current video frame.
  • the point coordinates and the direction vector can quickly and accurately achieve the azimuth tracking of the target entity, ensuring that the virtual globe displayed on the screen and the physical globe can move synchronously, giving the user a real sense of space presence and an actual operational experience.
  • FIG. 1 is a flowchart of a method for tracking an azimuth of a target entity according to an embodiment of the present application
  • FIG. 2( a ) is a schematic diagram of a feasible projection of a surface area of a globe in an equatorial plane provided by the present application.
  • FIG. 2(b) to (e) are schematic views showing the projection of the four hemispherical planes divided by the plane of Fig. 2(a) on the equatorial plane;
  • FIG. 3 is a schematic diagram showing the spatial orientation of two surface regions matching the visible surface in a reference coordinate system in the embodiment of the present application;
  • FIG. 5 is a flowchart of a method for implementing augmented reality according to an embodiment of the present application
  • FIG. 6 is a schematic structural diagram of an azimuth tracking device of a target entity according to an embodiment of the present application.
  • FIG. 7 is a schematic structural diagram of an apparatus for implementing augmented reality according to an embodiment of the present disclosure.
  • the terminal, the device of the service network, and the trusted party are both included.
  • One or more processors CPUs
  • input/output interfaces network interfaces
  • memory volatile and non-volatile memory
  • the memory may include non-persistent memory, random access memory (RAM), and/or non-volatile memory in a computer readable medium, such as read only memory (ROM) or flash memory.
  • RAM random access memory
  • ROM read only memory
  • Memory is an example of a computer readable medium.
  • Computer readable media includes both permanent and non-persistent, removable and non-removable media, and information storage can be implemented by any method or technology.
  • the information can be computer readable instructions, data structures, modules of programs, or other data.
  • Examples of storage media for a computer include, but are not limited to, phase change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), other types of random access memory (RAM), read only memory.
  • PRAM phase change memory
  • SRAM static random access memory
  • DRAM dynamic random access memory
  • RAM random access memory
  • ROM read only memory
  • EEPROM electrically erasable programmable read only memory
  • flash memory or other memory technology
  • compact disk read only memory CD-ROM
  • DVD digital versatile disk
  • a magnetic tape cartridge magnetic tape storage or other magnetic storage device or any other non-transportable medium that can be used to store information that can be accessed by a computing device.
  • computer readable media does not include non-transitory computer readable media, such as modulated data signals and carrier waves.
  • FIG. 1 is a schematic diagram of an azimuth tracking method for a target entity according to an embodiment of the present application.
  • the method is used to implement azimuth tracking of a visible surface of a physical target during processing of AR (Augmented Reality). Specifically, the following steps are included:
  • Step S101 Acquire a video frame including a target entity.
  • the target entity may be a physical globe, or may be other regular or irregularly shaped entities such as spheres, ellipsoids, polyhedra, and the like.
  • a target entity can be photographed by a device with an image capture device such as a camera to obtain a video frame containing the target entity.
  • an image capture device such as a camera
  • a video of a physical globe is taken by a camera of a smartphone, and any frame in the video can be used for subsequent steps to obtain a center coordinate point and a direction vector of the visible surface of the physical globe in the video frame.
  • Step S102 performing image feature extraction on the target entity in the video frame, and acquiring image feature information of the visible surface of the target entity.
  • a certain target entity only the surface of a certain part of the surface can be seen in the video frame.
  • the visible surface of the physical globe in the video frame is at most half a spherical surface, and when image feature extraction is performed, Also only need to mention Take the visual half-spherical image feature information.
  • the visible surface may include one or more of the sides depending on the angle of the camera lens.
  • image feature extraction various mature image feature extraction techniques in the prior art may be used, for example, based on any one or more combinations of image feature information such as color features, texture features, shape features, and spatial relationship features. Extracting one or more image feature information of the surface pattern of the globe.
  • Step S103 performing feature recognition on the acquired image feature information of the visible surface of the target entity based on the pre-made feature set of the target entity, and acquiring a surface area matching the visible surface in the pre-made feature set.
  • the prefabricated feature set includes a plurality of surface regions of the target entity and image feature information of each surface region
  • the physical globe is still taken as an example, and a plurality of surface regions included in the prefabricated feature set are feasible.
  • the division is as follows: dividing the globe by the direction of the warp, dividing the whole globe into four complete hemispheres, the center points of the four hemispheres (ie, the center of the sphere) are completely coincident, and each adjacent two hemispheres There is a 90° overlap between the two.
  • the projection of this division on the equatorial plane is shown in Fig. 2(a), and the projection of each hemisphere on the equatorial plane is shown in Fig. 2(b) to (e).
  • the preset feature set may directly adopt the existing set data, or may be obtained by offline pre-processing before processing.
  • the image feature information corresponding to the four hemispheres will also be completely different.
  • the image feature information corresponding to each hemispherical surface may also adopt a combination of any one or more of image feature information such as a color feature, a texture feature, a shape feature, and a spatial relationship feature.
  • the two can be matched, thereby obtaining the prefabricated feature. All surface areas in the collection that match the visible surface.
  • the visible surface of the acquired physical globe may contain a complete hemisphere in the set of prefabricated features, or may be composed of multiple A part of the hemisphere.
  • the lens direction coincides with any point on the equator and the geocentric line.
  • the visible surface of the physical globe may contain a complete hemisphere in the set of prefabricated features, or a part of two hemispheres; If the earth axes coincide, then the visible surface of the physical globe may be composed of portions of the four hemispheres in the set of prefabricated features described above.
  • Step S104 determining a center point coordinate and a direction vector of the visible surface of the target entity according to a center point coordinate and a direction vector of the surface area matched with the visible surface.
  • the position of the current camera is set as the origin of the reference coordinate system. In this reference coordinate system, whether the camera moves or the target entity moves, the position of the target entity can be regarded as occurring. The change, while the position of the camera remains the same. Since the spatial position of the visible surface of the target entity can be determined in the video frame, the center point coordinates and the direction vector of the surface area that matches the visible surface can thereby be determined.
  • the center point coordinate of the surface area is the coordinates of the center of the sphere in the reference coordinate system whose origin is the position where the camera is located, and the direction vector can be specified as the point from the center of the sphere under the same reference frame.
  • the unit vector to the longitude and latitude are 0° points.
  • the center point coordinate and the direction vector of the surface area are the center point coordinates and the direction vector of the visible surface of the target entity.
  • the center point coordinates and the direction vector of the surface area acquired by the video frame may exist due to factors unavoidable such as processing precision, camera shake, and the like. error.
  • the coordinates of the center points corresponding to the surface areas of the same target entity should be consistent, and in the presence of errors, the center point coordinates of the multiple surface areas may not be the same, and the corresponding direction vectors may also be different.
  • the spatial orientation of the two surface regions matching the visible surface in the reference coordinate system is as shown in FIG. 3, the coordinates of the center point are respectively with The direction vector is with
  • a center point coordinate of a visible surface of the target entity according to a center point coordinate of a surface area matching the visible surface, if there are multiple surface areas matching the visible surface, a plurality of surface areas
  • the center point coordinates are subjected to a weighted averaging process to determine the center point coordinates of the target entity.
  • the setting of the weights may be specifically set according to different scenarios to achieve an optimal calculation precision, and as a feasible implementation manner, the weights of all surface regions may be set. Set to the same, you can calculate according to the following formula:
  • the coordinates of the center point of the visible surface of the target entity Represents the center point coordinates of any surface area that matches the visible surface, where n is the number of matched surface areas.
  • the target entity visible surface direction vector is determined according to the direction vector of the surface area matching the visible surface, if there are multiple surface areas matching the visible surface, the direction of the plurality of surface areas
  • the vector weights are added and subjected to vector normalization processing to obtain a direction vector of the visible surface of the target entity.
  • the weight setting can be specifically set according to different scenarios to achieve optimal calculation accuracy, and as a feasible implementation manner, the weights of all surface areas can be set to be the same. It can be calculated according to the following formula:
  • the spatial orientation of the physical globe at each moment can be dynamically obtained. Whether changing the relative position of the camera and the physical globe, or rotating the physical globe, the physical globe can be accurately acquired relative to the center of the camera. The point coordinates and the direction vector, and based on this, the virtual globe is rendered to ensure that the virtual globe and the physical globe rotate or move in synchronization.
  • the target entity may be pre-processed before the video frame containing the target entity is acquired to construct a pre-made feature set of the target entity.
  • a corresponding set of prefabricated features can be constructed for a target entity of any shape, thereby realizing real-time azimuth tracking and AR implementation of the target entity of any shape.
  • the main process of the preprocessing is to scan the image features on the surface of the sphere offline to obtain a prefabricated feature set on the surface of the sphere.
  • the prefabricated feature set is pre-stored and can be collected in real time.
  • the image features of the visible surface of the physical globe are compared and matched.
  • the specific process includes the steps shown in Figure 4:
  • Step S401 acquiring an image of multiple surface areas of the target entity, wherein the plurality of surface areas cover at least all surface areas of the target entity;
  • Step S402 performing image feature extraction on the images of the plurality of surface regions, and acquiring image feature information of each surface region;
  • Step S403 construct a pre-made feature set of the target entity according to the image feature information of each surface area.
  • the prefabricated feature set is composed of a plurality of surface regions of the surface of the sphere (relative to the entire sphere surface, the surface region being a partial curved surface), and the coverage area of each surface region is no more than one complete hemisphere;
  • each two adjacent surface areas may have a certain overlap area
  • the embodiment of the present application further provides a method for implementing augmented reality.
  • the processing flow of the method is as shown in FIG. 5, and includes the following steps:
  • Step S501 Acquire a center point coordinate and a direction vector of the target entity by using the foregoing azimuth tracking method.
  • Step S502 obtaining a display portion corresponding to the visible surface from the virtual model according to the direction vector.
  • the virtual model may be a virtual globe that matches the physical globe.
  • a display portion corresponding to the visible surface may be determined based on a center of the virtual globe, and the display portion is a hemispherical surface pointed by the direction vector.
  • Step S503 rendering the display portion into the video frame according to the center point coordinate, so that a visible surface of the target entity is covered by the display portion.
  • the display portion Based on the coordinates of the center point, it can be determined at which position in the video frame the display portion needs to be rendered.
  • the hemispherical surface of the display portion can be determined in the video frame based on the coordinates of the center point and the radius of the virtual globe. The spatial position that should be in, thereby completing the synthesis of the final picture, such that the visible surface of the original physical globe in the video frame is covered by the corresponding hemisphere of the virtual globe.
  • Step S504 outputting the completed video frame.
  • the physical globe can not be seen in the picture presented to the user, but the virtual globe replaces the original position of the physical globe, because the virtual globe can display a large amount of information, and in the process of manipulating the physical globe, The virtual globe rotates at the same angle and speed, greatly improving the user experience.
  • virtual globes can be used to carry large amounts of information.
  • virtual information of different topics can be added, for example, in correspondence.
  • the geographical location presents three-dimensional models of landmark buildings, rare animals, and geological wonders.
  • the specific treatment method is:
  • an information superposition point in the display portion is acquired.
  • the display part contains a preset information superposition point, for example, an information superposition point about the giant panda is preset in the position of Sichuan, when the virtual globe is set.
  • the superposition point of the information can be detected.
  • the virtual information corresponding to the information superposition point is rendered into the video frame, so that the corresponding position of the display portion displays the virtual information.
  • the virtual information can be a three-dimensional model of a giant panda, so that the user can directly see a giant panda in Sichuan on the screen, and more intuitive and clear understanding of various geographical knowledge.
  • some trigger points can be set on the virtual model.
  • the trigger area can be set according to actual needs. For example, for a touch screen, you can The trigger area is set in the area currently touched by the user, so that the user clicks to implement the trigger; or can be statically set in a specific area of the screen, such as the center of the screen; and can be further marked by a cursor, an aperture, etc., to facilitate user identification. .
  • the triggering effect may be a voice, a video, a three-dimensional animation effect, etc., for example, setting the trigger point to a display position of the giant panda. When the three-dimensional model of the giant panda is in the triggering region, the three-dimensional model of the giant panda is Zoom in and perform corresponding actions at the same time, such as walking, standing, etc.
  • the triggering effect can also be a highlighting effect, for example, setting the trigger point to the range of any administrative area, when an administrative area enters the triggering area, the administrative The area will be highlighted.
  • a highlighting effect for example, setting the trigger point to the range of any administrative area, when an administrative area enters the triggering area, the administrative The area will be highlighted.
  • the triggering area, the triggering effect, and the setting manner of the triggering point are only examples, and other existing or future possible manners may be applied to the present application, and should also be included in the present application. It is within the scope of protection and is hereby incorporated by reference.
  • an orientation tracking device 60 for a target entity is also provided.
  • the device 60 is configured to implement azimuth tracking of a visible surface of a physical target during processing of implementing the AR.
  • 6 shows an image acquisition module 610, a feature extraction module 620, a feature matching module 630, and an integrated processing module 640.
  • the image collection module 610 is configured to acquire a video frame that includes a target entity.
  • the target entity may be a physical globe, or may be other regular or irregularly shaped entities such as spheres, ellipsoids, polyhedra, and the like.
  • a target entity can be photographed by a device with an image capture device such as a camera to obtain a video frame containing the target entity.
  • a video of a physical globe is taken by a camera of a smartphone, and any frame in the video can be used for subsequent processing to obtain a central coordinate point and a direction vector of the visible surface of the physical globe in the video frame.
  • the feature extraction module 620 is configured to perform image feature extraction on a target entity in the video frame to acquire image feature information of a visible surface of the target entity.
  • a target entity For a certain target entity, only the surface of a certain part of the surface can be seen in the video frame.
  • the visible surface of the physical globe in the video frame is at most half a spherical surface, and when image feature extraction is performed, It is also only necessary to extract visual half-spherical image feature information.
  • the visible surface may include one or more of the sides depending on the angle of the camera lens.
  • image feature extraction various mature image feature extraction techniques in the prior art can be used, for example, based on color features, texture features, and shapes.
  • image feature information such as signs and spatial relationship features, and extracting one or more image feature information of the surface pattern of the globe.
  • the feature matching module 630 is configured to perform feature recognition on the acquired image feature information of the target entity visible surface based on the set of pre-made features of the target entity, and acquire a matching match with the visible surface in the pre-made feature set.
  • Surface area Wherein the prefabricated feature set includes a plurality of surface regions of the target entity and image feature information of each surface region, and the physical globe is still taken as an example, and a plurality of surface regions included in the prefabricated feature set are feasible
  • the division is as follows: dividing the globe by the direction of the warp, dividing the whole globe into four complete hemispheres, the center points of the four hemispheres (ie, the center of the sphere) are completely coincident, and each adjacent two hemispheres There is a 90° overlap between the two.
  • the projection of this division on the equatorial plane is shown in Fig. 2(a), and the projection of each hemisphere on the equatorial plane is shown in Fig. 2(b) to (e).
  • the preset feature set may directly adopt the existing set data, or may be obtained by offline pre-processing before processing.
  • the image feature information corresponding to the four hemispheres will also be completely different.
  • the image feature information corresponding to each hemispherical surface may also adopt a combination of any one or more of image feature information such as a color feature, a texture feature, a shape feature, and a spatial relationship feature.
  • the two can be matched, thereby obtaining the prefabricated feature. All surface areas in the collection that match the visible surface.
  • the visible surface of the acquired physical globe may contain a complete hemisphere in the set of prefabricated features, or may be composed of multiple A part of the hemisphere.
  • the lens direction coincides with any point on the equator and the geocentric line
  • the visible surface of the physical globe may contain a complete hemisphere in the set of prefabricated features, or a part of two hemispheres;
  • the direction coincides with the earth axis, at which point the visible surface of the physical globe may consist of portions of the four hemispheres in the set of prefabricated features described above.
  • the comprehensive processing module 640 is configured to determine a center point coordinate and a direction vector of the visible surface of the target entity according to a center point coordinate and a direction vector of the surface area matched with the visible surface.
  • the position of the current camera is set as the origin of the reference coordinate system. In this reference coordinate system, whether the camera moves or the target entity moves, the position of the target entity can be regarded as occurring. The change, while the position of the camera remains the same. Since the spatial position of the visible surface of the target entity can be determined in the video frame, the center point coordinates and the direction vector of the surface area that matches the visible surface can thereby be determined.
  • the center point coordinate of the surface area is the coordinates of the center of the sphere in the reference coordinate system whose origin is the position where the camera is located, and the direction vector can be specified as the point from the center of the sphere under the same reference frame.
  • the unit vector to the longitude and latitude are 0° points.
  • the center point coordinate and the direction vector of the surface area are the center point coordinates and the direction vector of the visible surface of the target entity.
  • the center point coordinates and the direction vector of the surface area acquired by the video frame may exist due to factors unavoidable such as processing precision, camera shake, and the like. error.
  • the coordinates of the center points corresponding to the surface areas of the same target entity should be consistent, and in the presence of errors, the center point coordinates of the multiple surface areas may not be the same, and the corresponding direction vectors may also be different.
  • the spatial orientation of the two surface regions matching the visible surface in the reference coordinate system is as shown in FIG. 3, the coordinates of the center point are respectively with The direction vector is with
  • the integrated processing module 640 determines the center point coordinates of the visible surface of the target entity according to the center point coordinates of the surface area matching the visible surface, if there are multiple surface areas matching the visible surface And performing weighted averaging processing on center point coordinates of the plurality of surface regions to determine a center point coordinate of the target entity.
  • the weight setting may be specifically set according to different scenarios to achieve an optimal calculation precision, and as a feasible implementation manner, the weights of all surface regions may be set to be the same, and the calculation may be performed according to the following formula:
  • the coordinates of the center point of the visible surface of the target entity Represents the coordinates of the center point of any surface area that matches the visible surface, where n is the number of matched surface areas.
  • the integrated processing module 640 determines the target entity visible surface direction vector according to the direction vector of the surface area matching the visible surface, if there are multiple surface areas matching the visible surface, The direction vectors of the plurality of surface regions are weighted and added and subjected to vector normalization processing to obtain a direction vector of the visible surface of the target entity.
  • the weight setting can be specifically set according to different scenarios to achieve optimal calculation accuracy, and as a feasible implementation manner, the weights of all surface areas can be set to be the same. It can be calculated according to the following formula:
  • the spatial orientation of the physical globe at each moment can be dynamically obtained. Whether changing the relative position of the camera and the physical globe, or rotating the physical globe, the physical globe can be accurately acquired relative to the center of the camera. The point coordinates and the direction vector, and based on this, the virtual globe is rendered to ensure that the virtual globe and the physical globe rotate or move in synchronization.
  • the target entity may be pre-processed by the pre-processing module to obtain a pre-made feature set of the target entity before acquiring the video frame including the target entity.
  • a corresponding set of prefabricated features can be constructed for a target entity of any shape, thereby realizing real-time azimuth tracking and AR implementation of the target entity of any shape.
  • the main process of the preprocessing is to scan the image features on the surface of the sphere offline to obtain a prefabricated feature set on the surface of the sphere.
  • the prefabricated feature set is pre-stored and can be collected in real time.
  • the physical features of the physical globe are aligned and matched to the image features of the visible surface.
  • the pre-processing module is specifically configured to control the image acquisition module to acquire an image of a plurality of surface regions of the target entity, wherein the plurality of surface regions cover at least all surface regions of the target entity; and the feature extraction module is controlled Image of the plurality of surface regions is subjected to image feature extraction Taking image feature information of each surface area; and constructing a set of prefabricated features of the target entity according to image feature information of each surface area.
  • the prefabricated feature set is composed of a plurality of surface regions of the surface of the sphere (relative to the entire sphere surface, the surface region being a partial curved surface), and the coverage area of each surface region is no more than one complete hemisphere;
  • each two adjacent surface areas may have a certain overlap area
  • the embodiment of the present application further provides an apparatus for implementing augmented reality.
  • the structure of the apparatus is as shown in FIG. 7, and includes an azimuth tracking device 60, a rendering device 710, and an output device 720.
  • the azimuth tracking device 60 is configured to acquire a center point coordinate and a direction vector of the target entity.
  • the rendering device 710 is configured to acquire a display portion corresponding to the visible surface from a virtual model according to the direction vector, and render the display portion into the video frame according to the center point coordinate, so that The visible surface of the target entity is covered by the display portion.
  • the virtual model may be a virtual globe that matches the physical globe.
  • a display portion corresponding to the visible surface may be determined based on a center of the virtual globe, and the display portion is a hemispherical surface pointed by the direction vector. Based on the coordinates of the center point, it can be determined at which position in the video frame the display portion needs to be rendered. Still taking the virtual globe as an example, the hemispherical surface of the display portion can be determined in the video frame based on the coordinates of the center point and the radius of the virtual globe. The spatial position that should be in, thereby completing the synthesis of the final picture, such that the visible surface of the original physical globe in the video frame is covered by the corresponding hemisphere of the virtual globe.
  • the output device 720 is configured to output a video frame that is finished rendering.
  • the output device 720 is configured to output a video frame that is finished rendering.
  • the device may include, but is not limited to, a user terminal or a device formed by integrating a user terminal and a network device through a network.
  • the user terminal includes, but is not limited to, a personal computer, a touch terminal, and the like, and specifically may be a smart phone, a tablet computer, a PDA, an AR glasses, or other electronic device having an image acquisition, processing, and output function; Implementations are not limited to, for example, a network host, a single network server, multiple network server sets, or a cloud-based computer collection.
  • the cloud is composed of a large number of host or network servers based on Cloud Computing, which is a kind of distributed computing, a virtual computer composed of a group of loosely coupled computers.
  • the device may also be an electronic device running an application (APP) containing associated algorithms and graphical user interfaces, or the application itself.
  • APP application
  • the camera of the smart phone can be used to implement the related functions of the image capturing module 610
  • the processor can be used to implement the feature extraction module 620, the feature matching module 630, the comprehensive processing module 640, and
  • the related functions of the device 710 are rendered while their screens enable the associated functions of the output device 720.
  • the processor can also upload relevant data to the network device through the communication module, and the network device completes calculation and processing of related data.
  • virtual globes can be used to carry large amounts of information.
  • virtual information of different topics can be added, for example, in correspondence.
  • the geographical location presents three-dimensional models of landmark buildings, rare animals, and geological wonders.
  • the rendering device 710 is further configured to perform the following processing:
  • an information superposition point in the display portion is acquired.
  • the display part contains a pre-set information superposition point, for example, a superposition point about the giant panda is preset in the position of Sichuan, when the virtual globe is set.
  • a superposition point about the giant panda is preset in the position of Sichuan, when the virtual globe is set.
  • the hemisphere of the display part contains the geographical area where Sichuan is located, it can be detected. Go to the superposition point of this information.
  • the virtual information corresponding to the information superposition point is rendered into the video frame, so that the corresponding position of the display portion displays the virtual information.
  • the virtual information can be a three-dimensional model of a giant panda, so that the user can directly see a giant panda in Sichuan on the screen, and more intuitive and clear understanding of various geographical knowledge.
  • some trigger points can be set on the virtual model.
  • the rendering device 710 determines whether the trigger point of the display portion is located in a trigger region of the video frame; if the trigger point is in a trigger region of the video frame, triggering the trigger point Rendered into the video frame.
  • the trigger area can be set according to actual needs. For example, for a touch screen, the trigger area may be set to an area currently touched by the user to implement a user click to implement triggering; or may be statically set in a certain specific area of the screen, such as the center of the screen; and may further pass a cursor, an aperture, etc. Mark it for user identification.
  • the triggering effect may be audio, video, three-dimensional animation effect, etc., for example, setting the trigger point to the display position of the giant panda. When the three-dimensional model of the giant panda is in the triggering area, the three-dimensional model of the giant panda will be Zoom in and perform corresponding actions at the same time, such as walking, standing, etc.
  • the triggering effect can also be a highlighting effect, for example, setting the trigger point to the range of any administrative area, when an administrative area enters the triggering area, the administrative The area will be highlighted.
  • a highlighting effect for example, setting the trigger point to the range of any administrative area, when an administrative area enters the triggering area, the administrative The area will be highlighted.
  • the triggering area, the triggering effect, and the setting manner of the triggering point are only examples, and other existing or future possible manners may be applied to the present application, and should also be included in the present application. It is within the scope of protection and is hereby incorporated by reference.
  • the solution of the present application can superimpose some virtual graphics, images, information, etc. into a real scene based on augmented reality technology, and enhances the real scene by using the physical globe as a target entity. It combines the advantages of both physical globes and virtual globes in the prior art.
  • Azimuth tracking is realized by acquiring the central coordinate point and direction vector of the target entity in real time, and the output video frame is rendered in real time based on the central coordinate point and the direction vector, so that the virtual globe will be the same in the process of manipulating the physical globe.
  • Angle and speed rotation can greatly improve the user experience.
  • image feature extraction is performed by a video frame containing the target entity acquired in real time
  • Image feature matching is performed with the pre-built pre-made feature set, and calculation is performed based on the matching result to obtain the center point coordinate and the direction vector of the visible surface of the target entity in the current video frame, thereby enabling fast and accurate orientation tracking of the target entity.
  • the present application can be implemented in software and/or a combination of software and hardware, for example, using an application specific integrated circuit (ASIC), a general purpose computer, or any other similar hardware device.
  • the software program of the present application can be executed by a processor to implement the steps or functions described above.
  • the software programs (including related data structures) of the present application can be stored in a computer readable recording medium such as a RAM memory, a magnetic or optical drive or a floppy disk and the like.
  • some of the steps or functions of the present application may be implemented in hardware, for example, as a circuit that cooperates with a processor to perform various steps or functions.
  • a portion of the present application can be applied as a computer program product, such as computer program instructions, which, when executed by a computer, can invoke or provide a method and/or technical solution in accordance with the present application.
  • the program instructions for invoking the method of the present application may be stored in a fixed or removable recording medium, and/or transmitted by a data stream in a broadcast or other signal bearing medium, and/or stored in a The working memory of the computer device in which the program instructions are run.
  • an embodiment in accordance with the present application includes a device including a memory for storing computer program instructions and a processor for executing program instructions, wherein when the computer program instructions are executed by the processor, triggering
  • the apparatus operates based on the aforementioned methods and/or technical solutions in accordance with various embodiments of the present application.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Computer Hardware Design (AREA)
  • Computer Graphics (AREA)
  • Astronomy & Astrophysics (AREA)
  • Business, Economics & Management (AREA)
  • Educational Administration (AREA)
  • Educational Technology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The present application aims to provide a method and a relevant apparatus for orientational tracking, and a method and a device for realizing augmented reality, which, on the basis of augmented reality technology, can superpose some virtual graphics, images, information, etc., into a real scene, augmenting the real scene. The invention takes a physical globe as a target entity, and enables the physical globe to possess the advantages of both a physical globe in the prior art and a virtual globe. The present invention realizes orientational tracking by acquiring, in real time, the central coordinate point and a direction vector of a target entity, and renders, in real time, on the basis of the central coordinate point and direction vector, outputted video frames, so that when a user manipulates a physical globe, a virtual globe will rotate by the same angle and speed, greatly improving user experience.

Description

方位跟踪方法、实现增强现实的方法及相关装置、设备Azimuth tracking method, method for realizing augmented reality, and related device and device 技术领域Technical field
本申请涉及计算机领域,尤其涉及一种方位跟踪方法、实现增强现实的方法及相关装置、设备。The present application relates to the field of computers, and in particular, to an azimuth tracking method, a method for implementing augmented reality, and related devices and devices.
背景技术Background technique
在人们认识世界的过程中,地球仪一直是不可缺少的地理教具,它可以让人们以更加形象而有趣的方式来认识空间与位置的概念,了解到陆地的宽广、海洋的浩瀚、四季的更迭,尤其对于孩子们形成正确的科学观和世界观也具有重要的意义。但是,长期以来,传统的地球仪只能表达非常有限的地理信息,例如陆地与海洋的边界、国家之间的边境线等;而对于更加丰富的地理信息,例如地球内部的构造、地球在太阳系中的位置、地球表面的动植物分布规律等,传统的地球仪则无能为力。In the process of people knowing the world, the globe has always been an indispensable geographic teaching aid. It allows people to understand the concept of space and location in a more vivid and interesting way, and to understand the vastness of the land, the vastness of the ocean, and the change of the four seasons. Especially for children to form a correct scientific and worldview is also of great significance. However, for a long time, traditional globes can only express very limited geographic information, such as the boundary between land and ocean, the border between countries, etc., and for richer geographic information, such as the internal structure of the Earth, the Earth in the solar system. The position of the earth, the distribution of animals and plants on the surface of the earth, etc., the traditional globe is powerless.
目前存在一些虚拟地球仪应用产品,此类应用产品通过屏幕显示一个虚拟地球仪,人们可以通过单指在屏幕上滑动来控制虚拟地球仪转动,也可以通过双指在屏幕上伸缩来控制虚拟地球仪放大或缩小。相对于实体地球仪而言,虚拟地球仪操作简便、便于携带,信息量也很大。但是,虚拟地球仪也存在一些明显的劣势,例如用户体验相对较差,无法给用户以真实的空间存在感和实际的操作体验感。There are some virtual globe application products. These applications display a virtual globe through the screen. People can control the rotation of the virtual globe by sliding on the screen with a single finger. It can also control the virtual globe to zoom in or out by scaling the two fingers on the screen. . Compared to physical globes, virtual globes are easy to operate, easy to carry, and have a large amount of information. However, virtual globes also have some obvious disadvantages, such as the relatively poor user experience, which can not give users a real sense of space and actual operational experience.
申请内容Application content
本申请的目的是提供一种方位跟踪方法、实现增强现实的方法及相关装置、设备。The purpose of the application is to provide an orientation tracking method, a method for implementing augmented reality, and related devices and devices.
为实现上述目的,本申请提供了一种目标实体的方位跟踪方法,该方法包括:To achieve the above objective, the present application provides a method for tracking an orientation of a target entity, the method comprising:
获取包含目标实体的视频帧;Get a video frame containing the target entity;
对所述视频帧中的目标实体进行图像特征提取,获取所述目标实体可视表面的图像特征信息; Performing image feature extraction on the target entity in the video frame to acquire image feature information of the visible surface of the target entity;
基于所述目标实体的预制特征集合对获取到的所述目标实体可视表面的图像特征信息进行特征识别,在预制特征集合中获取与所述可视表面匹配的表面区域,其中,所述预制特征集合包含所述目标实体的多个表面区域以及每个表面区域的图像特征信息;Performing feature recognition on the acquired image feature information of the visible surface of the target entity based on the set of prefabricated features of the target entity, and acquiring a surface area matching the visible surface in the set of prefabricated features, wherein the prefabrication The feature set includes a plurality of surface regions of the target entity and image feature information of each surface region;
根据与所述可视表面匹配的表面区域的中心点坐标和方向矢量分别确定所述目标实体可视表面的中心点坐标和方向矢量。A center point coordinate and a direction vector of the visible surface of the target entity are respectively determined according to a center point coordinate and a direction vector of the surface area matching the visible surface.
进一步地,在获取包含目标实体的视频帧之前,还包括:Further, before acquiring the video frame including the target entity, the method further includes:
对目标实体进行预处理,构建所述目标实体的预制特征集合。Preprocessing the target entity to construct a set of prefabricated features of the target entity.
进一步地,对目标实体进行预处理,构建所述目标实体的预制特征集合,包括:Further, the target entity is preprocessed to construct a prefabricated feature set of the target entity, including:
获取目标实体多个表面区域的图像,其中,所述多个表面区域至少覆盖所述目标实体的所有表面区域;Acquiring an image of a plurality of surface regions of the target entity, wherein the plurality of surface regions cover at least all surface regions of the target entity;
对所述多个表面区域的图像进行图像特征提取,获取所述每个表面区域的图像特征信息;Performing image feature extraction on the images of the plurality of surface regions to acquire image feature information of each surface region;
根据所述每个表面区域的图像特征信息构建所述目标实体的预制特征集合。Constructing a set of prefabricated features of the target entity based on the image feature information of each of the surface regions.
进一步地,根据与所述可视表面匹配的表面区域的中心点坐标确定所述目标实体可视表面的中心点坐标,包括:Further, determining a center point coordinate of the visible surface of the target entity according to a center point coordinate of the surface area matching the visible surface comprises:
若与所述可视表面匹配的表面区域有多个,将多个表面区域的中心点坐标进行加权平均处理,确定所述目标实体的中心点坐标。If there are a plurality of surface regions matching the visible surface, the center point coordinates of the plurality of surface regions are subjected to weighted averaging processing, and the center point coordinates of the target entity are determined.
进一步地,根据与所述可视表面匹配的表面区域的方向矢量确定所述目标实体可视表面方向矢量,包括:Further, determining the target entity visible surface direction vector according to a direction vector of the surface area matching the visible surface comprises:
若与所述可视表面匹配的表面区域有多个,将多个表面区域的方向矢量加权相加并进行矢量归一化处理,获取所述目标实体可视表面的方向矢量。If there are multiple surface regions matching the visible surface, the direction vectors of the plurality of surface regions are weighted and added and subjected to vector normalization processing to obtain a direction vector of the visible surface of the target entity.
本申请还提供了一种实现增强现实的方法,该方法包括:The application also provides a method for implementing augmented reality, the method comprising:
采用所述方位跟踪方法获取所述目标实体的中心点坐标和方向矢量;Obtaining a center point coordinate and a direction vector of the target entity by using the azimuth tracking method;
根据所述方向矢量由虚拟模型中获取与所述可视表面对应的显示部分; Obtaining a display portion corresponding to the visible surface from the virtual model according to the direction vector;
根据所述中心点坐标将所述显示部分渲染至所述视频帧中,以使所述目标实体的可视表面由所述显示部分覆盖。The display portion is rendered into the video frame based on the center point coordinates such that a visible surface of the target entity is covered by the display portion.
输出完成渲染的视频帧。Output the finished rendered video frame.
进一步地,该方法还包括:Further, the method further includes:
获取所述显示部分中的信息叠加点;Obtaining an information superposition point in the display portion;
将所述信息叠加点对应的虚拟信息渲染至所述视频帧中,以使所述显示部分的相应位置显示所述虚拟信息。And rendering virtual information corresponding to the information superposition point into the video frame, so that the corresponding position of the display part displays the virtual information.
进一步地,所述虚拟模型设置有触发点;Further, the virtual model is provided with a trigger point;
所述方法还包括:The method further includes:
判断所述显示部分的触发点是否位于所述视频帧的触发区域;Determining whether a trigger point of the display portion is located in a trigger area of the video frame;
若所述触发点处于所述视频帧的触发区域,将所述触发点的触发效果渲染至所述视频帧中。If the trigger point is in a trigger area of the video frame, the trigger effect of the trigger point is rendered into the video frame.
进一步地,所述目标实体为实体地球仪,所述虚拟模型为与所述实体地球仪匹配的虚拟地球仪。Further, the target entity is a physical globe, and the virtual model is a virtual globe matching the physical globe.
基于本申请的另一方面,还提供了一种目标实体的方位跟踪装置,该装置包括:According to another aspect of the present application, there is also provided an orientation tracking device for a target entity, the device comprising:
图像采集模块,用于获取包含目标实体的视频帧;An image acquisition module, configured to acquire a video frame that includes a target entity;
特征提取模块,用于对所述视频帧中的目标实体进行图像特征提取,获取所述目标实体可视表面的图像特征信息;a feature extraction module, configured to perform image feature extraction on a target entity in the video frame, and acquire image feature information of a visible surface of the target entity;
特征匹配模块,用于基于所述目标实体的预制特征集合对获取到的所述目标实体可视表面的图像特征信息进行特征识别,在预制特征集合中获取与所述可视表面匹配的表面区域,其中,所述预制特征集合包含所述目标实体的多个表面区域以及每个表面区域的图像特征信息;a feature matching module, configured to perform feature recognition on the acquired image feature information of the visible surface of the target entity based on the set of prefabricated features of the target entity, and acquire a surface region matching the visible surface in the prefabricated feature set The set of prefabricated features includes a plurality of surface regions of the target entity and image feature information of each surface region;
综合处理模块,用于根据与所述可视表面匹配的表面区域的中心点坐标和方向矢量分别确定所述目标实体可视表面的中心点坐标和方向矢量。And an integrated processing module, configured to respectively determine a center point coordinate and a direction vector of the visible surface of the target entity according to a center point coordinate and a direction vector of the surface area matching the visible surface.
进一步地,该装置还包括:Further, the device further includes:
预处理模块,用于在获取包含目标实体的视频帧之前,对目标实体进行预处理,构建所述目标实体的预制特征集合。And a pre-processing module, configured to pre-process the target entity before acquiring the video frame that includes the target entity, and construct a pre-made feature set of the target entity.
进一步地,所述预处理模块,用于控制所述图像采集模块获取目标实 体多个表面区域的图像,其中,所述多个表面区域至少覆盖所述目标实体的所有表面区域;控制所述特征提取模块对所述多个表面区域的图像进行图像特征提取,获取所述每个表面区域的图像特征信息;以及根据所述每个表面区域的图像特征信息构建所述目标实体的预制特征集合。Further, the preprocessing module is configured to control the image acquisition module to acquire a target real An image of a plurality of surface regions, wherein the plurality of surface regions cover at least all surface regions of the target entity; controlling the feature extraction module to perform image feature extraction on images of the plurality of surface regions to obtain the Image feature information for each surface region; and constructing a set of prefabricated features of the target entity based on image feature information for each of the surface regions.
进一步地,所述综合处理模块,用于在与所述可视表面匹配的表面区域有多个时,将多个表面区域的中心点坐标进行加权平均处理,确定所述目标实体的中心点坐标。Further, the integrated processing module is configured to perform weighted averaging processing on center point coordinates of the plurality of surface regions when there are multiple surface regions matching the visible surface, and determine a center point coordinate of the target entity .
进一步地,所述综合处理模块,用于在与所述可视表面匹配的表面区域有多个时,将多个表面区域的方向矢量加权相加并进行矢量归一化处理,获取所述目标实体可视表面的方向矢量。Further, the integrated processing module is configured to weight-add the direction vectors of the plurality of surface regions and perform vector normalization processing to obtain the target when there are multiple surface regions matching the visible surface The direction vector of the solid visible surface.
本申请还提供了一种实现增强现实的设备,该设备包括:The application also provides a device for realizing augmented reality, the device comprising:
方位跟踪装置,用于获取所述目标实体的中心点坐标和方向矢量;An orientation tracking device, configured to acquire a center point coordinate and a direction vector of the target entity;
渲染装置,用于根据所述方向矢量由虚拟模型中获取与所述可视表面对应的显示部分,以及根据所述中心点坐标将所述显示部分渲染至所述视频帧中,以使所述目标实体的可视表面由所述显示部分覆盖;a rendering device, configured to acquire a display portion corresponding to the visible surface from the virtual model according to the direction vector, and render the display portion into the video frame according to the center point coordinate, so that the The visible surface of the target entity is covered by the display portion;
输出装置,用于输出完成渲染的视频帧。An output device for outputting a finished video frame.
进一步地,所述渲染装置,还用于获取所述显示部分中的信息叠加点,以及将所述信息叠加点对应的虚拟信息渲染至所述视频帧中,以使所述显示部分的相应位置显示所述虚拟信息。Further, the rendering device is further configured to acquire an information superposition point in the display portion, and render virtual information corresponding to the information superimposition point into the video frame to make a corresponding position of the display portion The virtual information is displayed.
进一步地,所述虚拟模型设置有触发点;Further, the virtual model is provided with a trigger point;
所述渲染装置,还用于判断所述显示部分的触发点是否位于所述视频帧的触发区域,并在所述触发点处于所述视频帧的触发区域时,将所述触发点的触发效果渲染至所述视频帧中。The rendering device is further configured to determine whether a trigger point of the display portion is located in a trigger region of the video frame, and when the trigger point is in a trigger region of the video frame, triggering the trigger point Rendered into the video frame.
进一步地,所述目标实体为实体地球仪,所述虚拟模型为与所述实体地球仪匹配的虚拟地球仪。Further, the target entity is a physical globe, and the virtual model is a virtual globe matching the physical globe.
与现有技术相比,本申请的方案能够基于增强现实技术将一些虚拟的图形、图像、信息等叠加到真实场景中去,起到了对现实场景进行增强的作用,通过将实体地球仪作为目标实体,使其能够同时兼备现有技术中的实体地球仪和虚拟地球仪的优点。通过实时获取目标实体的中心坐标点和 方向矢量实现方位跟踪,并基于该中心坐标点和方向矢量对输出的视频帧进行实时渲染,使得用户在操控实体地球仪的过程中,虚拟地球仪会以相同的角度和速度转动,能够极大地改善用户体验。Compared with the prior art, the solution of the present application can superimpose some virtual graphics, images, information, etc. into a real scene based on augmented reality technology, and enhances the real scene by using the physical globe as a target entity. It enables it to combine the advantages of both physical globes and virtual globes in the prior art. Get the central coordinate point of the target entity in real time and The direction vector realizes the azimuth tracking, and the real-time rendering of the output video frame based on the central coordinate point and the direction vector, so that the virtual globe rotates at the same angle and speed during the manipulation of the physical globe, which can greatly improve the user. Experience.
此外,通过实时获取的包含目标实体的视频帧,进行图像特征提取,并与预先构建的预制特征集合进行图像特征匹配,并基于匹配结果进行计算,获取当前视频帧中目标实体可视表面的中心点坐标和方向矢量,由此能够快速准确地实现目标实体的方位跟踪,保证屏幕中显示的虚拟地球仪与实体地球仪能够同步运动,给予用户真实的空间存在感和实际的操作体验感。In addition, the image feature extraction is performed by the video frame containing the target entity acquired in real time, and the image feature matching is performed with the pre-built pre-made feature set, and the calculation is performed based on the matching result to obtain the center of the visible surface of the target entity in the current video frame. The point coordinates and the direction vector can quickly and accurately achieve the azimuth tracking of the target entity, ensuring that the virtual globe displayed on the screen and the physical globe can move synchronously, giving the user a real sense of space presence and an actual operational experience.
附图说明DRAWINGS
通过阅读参照以下附图所作的对非限制性实施例所作的详细描述,本申请的其它特征、目的和优点将会变得更明显:Other features, objects, and advantages of the present application will become more apparent from the detailed description of the accompanying drawings.
图1为本申请实施例提供的一种目标实体的方位跟踪方法的流程图;FIG. 1 is a flowchart of a method for tracking an azimuth of a target entity according to an embodiment of the present application;
图2(a)为本申请提供的一种可行的地球仪表面区域划分方式在赤道平面的投影示意图FIG. 2( a ) is a schematic diagram of a feasible projection of a surface area of a globe in an equatorial plane provided by the present application.
图2(b)~(e)分别为图2(a)所划分的4个半球面在赤道平面的投影示意图;2(b) to (e) are schematic views showing the projection of the four hemispherical planes divided by the plane of Fig. 2(a) on the equatorial plane;
图3为本申请实施例中与所述可视表面匹配的两个表面区域在参考坐标系下的空间方位示意图;3 is a schematic diagram showing the spatial orientation of two surface regions matching the visible surface in a reference coordinate system in the embodiment of the present application;
图4为本申请实施例中进行预处理的具体处理流程图;4 is a specific processing flowchart of performing preprocessing in the embodiment of the present application;
图5为本申请实施例提供的一种实现增强现实的方法的流程图;FIG. 5 is a flowchart of a method for implementing augmented reality according to an embodiment of the present application;
图6为申请实施例提供的一种目标实体的方位跟踪装置的结构示意图;6 is a schematic structural diagram of an azimuth tracking device of a target entity according to an embodiment of the present application;
图7为本申请实施例提供的一种实现增强现实的设备的结构示意图;FIG. 7 is a schematic structural diagram of an apparatus for implementing augmented reality according to an embodiment of the present disclosure;
附图中相同或相似的附图标记代表相同或相似的部件。The same or similar reference numerals in the drawings denote the same or similar components.
具体实施方式detailed description
下面结合附图对本申请作进一步详细描述。The present application is further described in detail below with reference to the accompanying drawings.
在本申请一个典型的配置中,终端、服务网络的设备和可信方均包括 一个或多个处理器(CPU)、输入/输出接口、网络接口和内存。In a typical configuration of the present application, the terminal, the device of the service network, and the trusted party are both included. One or more processors (CPUs), input/output interfaces, network interfaces, and memory.
内存可能包括计算机可读介质中的非永久性存储器,随机存取存储器(RAM)和/或非易失性内存等形式,如只读存储器(ROM)或闪存(flash RAM)。内存是计算机可读介质的示例。The memory may include non-persistent memory, random access memory (RAM), and/or non-volatile memory in a computer readable medium, such as read only memory (ROM) or flash memory. Memory is an example of a computer readable medium.
计算机可读介质包括永久性和非永久性、可移动和非可移动媒体,可以由任何方法或技术来实现信息存储。信息可以是计算机可读指令、数据结构、程序的模块或其他数据。计算机的存储介质的例子包括、但不限于相变内存(PRAM)、静态随机存取存储器(SRAM)、动态随机存取存储器(DRAM)、其他类型的随机存取存储器(RAM)、只读存储器(ROM)、电可擦除可编程只读存储器(EEPROM)、快闪记忆体或其他内存技术、只读光盘只读存储器(CD-ROM)、数字多功能光盘(DVD)或其他光学存储、磁盒式磁带,磁带磁盘存储或其他磁性存储设备或任何其他非传输介质,可用于存储可以被计算设备访问的信息。按照本文中的界定,计算机可读介质不包括非暂存电脑可读媒体(transitory media),如调制的数据信号和载波。Computer readable media includes both permanent and non-persistent, removable and non-removable media, and information storage can be implemented by any method or technology. The information can be computer readable instructions, data structures, modules of programs, or other data. Examples of storage media for a computer include, but are not limited to, phase change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), other types of random access memory (RAM), read only memory. (ROM), electrically erasable programmable read only memory (EEPROM), flash memory or other memory technology, compact disk read only memory (CD-ROM), digital versatile disk (DVD) or other optical storage, A magnetic tape cartridge, magnetic tape storage or other magnetic storage device or any other non-transportable medium that can be used to store information that can be accessed by a computing device. As defined herein, computer readable media does not include non-transitory computer readable media, such as modulated data signals and carrier waves.
图1示出了本申请实施例提供的一种目标实体的方位跟踪方法,该方法用于在实现AR(Augmented Reality,增强现实)的处理过程中,实现对实体目标可视表面的方位跟踪,具体包括以下步骤:FIG. 1 is a schematic diagram of an azimuth tracking method for a target entity according to an embodiment of the present application. The method is used to implement azimuth tracking of a visible surface of a physical target during processing of AR (Augmented Reality). Specifically, the following steps are included:
步骤S101,获取包含目标实体的视频帧。所述目标实体可以是实体地球仪,也可以是例如球体、椭球体、多面体等其它规则或者不规则形状的实体。在实际场景中可以通过带有摄像头等图像采集装置的设备来拍摄目标实体,以获取到包含目标实体的视频帧。例如,通过智能手机的摄像头拍摄一段关于实体地球仪的视频,该视频中的任意一帧均可用于后续步骤的处理,以获取在该视频帧中实体地球仪可视表面的中心坐标点和方向矢量。Step S101: Acquire a video frame including a target entity. The target entity may be a physical globe, or may be other regular or irregularly shaped entities such as spheres, ellipsoids, polyhedra, and the like. In an actual scenario, a target entity can be photographed by a device with an image capture device such as a camera to obtain a video frame containing the target entity. For example, a video of a physical globe is taken by a camera of a smartphone, and any frame in the video can be used for subsequent steps to obtain a center coordinate point and a direction vector of the visible surface of the physical globe in the video frame.
步骤S102,对所述视频帧中的目标实体进行图像特征提取,获取所述目标实体可视表面的图像特征信息。对于某一目标实体,在视频帧中仅可以看到其某一部分的表面,例如对于球形的实体地球仪,视频帧中该实体地球仪的可视表面最大为半个球面,在进行图像特征提取时,也仅需要提 取可视的半个球面的图像特征信息。而对于例如多面体或者其它不规则形状的实体,可视表面根据摄像头镜头角度的不同,可能包含其中的一个或者多个侧面。在进行图像特征提取时,可以采用现有技术中各类成熟的图像特征提取技术,例如基于颜色特征、纹理特征、形状特征、空间关系特征等图像特征信息中任意一种或多种的组合,提取地球仪表面图案的一种或多种图像特征信息。Step S102, performing image feature extraction on the target entity in the video frame, and acquiring image feature information of the visible surface of the target entity. For a certain target entity, only the surface of a certain part of the surface can be seen in the video frame. For example, for a spherical physical globe, the visible surface of the physical globe in the video frame is at most half a spherical surface, and when image feature extraction is performed, Also only need to mention Take the visual half-spherical image feature information. For entities such as polyhedrons or other irregular shapes, the visible surface may include one or more of the sides depending on the angle of the camera lens. In the image feature extraction, various mature image feature extraction techniques in the prior art may be used, for example, based on any one or more combinations of image feature information such as color features, texture features, shape features, and spatial relationship features. Extracting one or more image feature information of the surface pattern of the globe.
步骤S103,基于所述目标实体的预制特征集合对获取到的所述目标实体可视表面的图像特征信息进行特征识别,在预制特征集合中获取与所述可视表面匹配的表面区域。其中,所述预制特征集合包含所述目标实体的多个表面区域以及每个表面区域的图像特征信息,仍以前述实体地球仪为例,其预制特征集合中包含的多个表面区域的一种可行的划分方式为:以地球仪的经线方向进行划分,将整个地球仪划分为4个完整的半球面,这4个半球面的中心点(即球心)完全重合,且相邻的每两个半球面之间具有90°的重合区域,这种划分方式在赤道平面上的投影如图2(a)所示,其中各个半球面在赤道平面上的投影如图2(b)~(e)。所述预制特征集合可以直接采用已有的集合数据,也可以在进行处理之前,通过离线预处理的方式获取。Step S103, performing feature recognition on the acquired image feature information of the visible surface of the target entity based on the pre-made feature set of the target entity, and acquiring a surface area matching the visible surface in the pre-made feature set. Wherein the prefabricated feature set includes a plurality of surface regions of the target entity and image feature information of each surface region, and the physical globe is still taken as an example, and a plurality of surface regions included in the prefabricated feature set are feasible The division is as follows: dividing the globe by the direction of the warp, dividing the whole globe into four complete hemispheres, the center points of the four hemispheres (ie, the center of the sphere) are completely coincident, and each adjacent two hemispheres There is a 90° overlap between the two. The projection of this division on the equatorial plane is shown in Fig. 2(a), and the projection of each hemisphere on the equatorial plane is shown in Fig. 2(b) to (e). The preset feature set may directly adopt the existing set data, or may be obtained by offline pre-processing before processing.
而由于每个半球面上的地理环境不同,相应部分的图案也会各不相同,例如水域和陆地会采用不同的色彩,陆地中各种类型的地形同样会采用不同的颜色或图案等,由此4个半球面对应的图像特征信息也将完全不同。所述预制特征集合中,每个半球面对应的图像特征信息同样可以采用颜色特征、纹理特征、形状特征、空间关系特征等图像特征信息中任意一种或多种的组合。通过图像特征的匹配识别,若可视表面某一部分的图像特征信息与预制特征集合中的某一表面区域中的相应部分相同或者相似度超过阈值,则可以确定两者匹配,由此获取预制特征集合中所有与可视表面匹配的表面区域。Because the geographical environment on each hemisphere is different, the patterns of the corresponding parts will be different. For example, the water and land will adopt different colors, and various types of terrain on the land will also adopt different colors or patterns. The image feature information corresponding to the four hemispheres will also be completely different. In the prefabricated feature set, the image feature information corresponding to each hemispherical surface may also adopt a combination of any one or more of image feature information such as a color feature, a texture feature, a shape feature, and a spatial relationship feature. By matching the image features, if the image feature information of a certain part of the visible surface is the same as the corresponding part of a certain surface area in the prefabricated feature set or the similarity exceeds the threshold, the two can be matched, thereby obtaining the prefabricated feature. All surface areas in the collection that match the visible surface.
在实际场景中,由于在视频获取过程中摄像头与实体地球仪的相对位置的不同,获取到的实体地球仪的可视表面可能包含了上述预制特征集合中一个完整的半球面,也可能由其中多个半球面的某一部分组成。例如, 镜头方向与赤道上任意一点和地心连线重合,此时实体地球仪的可视表面可能包含上述预制特征集合中一个完整的半球面,或者由其中两个半球面的一部分组成;若镜头方向与地轴重合,则此时实体地球仪的可视表面可能由上述预制特征集合中4个半球面的各一部分组成。In an actual scenario, due to the difference in the relative positions of the camera and the physical globe during the video acquisition process, the visible surface of the acquired physical globe may contain a complete hemisphere in the set of prefabricated features, or may be composed of multiple A part of the hemisphere. E.g, The lens direction coincides with any point on the equator and the geocentric line. At this point, the visible surface of the physical globe may contain a complete hemisphere in the set of prefabricated features, or a part of two hemispheres; If the earth axes coincide, then the visible surface of the physical globe may be composed of portions of the four hemispheres in the set of prefabricated features described above.
步骤S104,根据与所述可视表面匹配的表面区域的中心点坐标和方向矢量分别确定所述目标实体可视表面的中心点坐标和方向矢量。本实施例在进行数学处理的过程中,将当前摄像头所在的位置设置为参考坐标系的原点,在此参考坐标系下,不论是摄像头移动还是目标实体移动,都可以视为目标实体的位置发生了变化,而摄像头的位置保持不变。由于目标实体可视表面的空间位置可以在视频帧中确定,则由此可以确定与所述可视表面匹配的表面区域的中心点坐标和方向矢量。对于实体地球仪,其表面区域的所述中心点坐标即为以摄像头所在的位置为原点的参考坐标系下球心的坐标,而所述方向矢量则可指定为在同样参考系下从球心点到经度、纬度均为0°点的单位矢量。Step S104, determining a center point coordinate and a direction vector of the visible surface of the target entity according to a center point coordinate and a direction vector of the surface area matched with the visible surface. In the process of performing mathematical processing, the position of the current camera is set as the origin of the reference coordinate system. In this reference coordinate system, whether the camera moves or the target entity moves, the position of the target entity can be regarded as occurring. The change, while the position of the camera remains the same. Since the spatial position of the visible surface of the target entity can be determined in the video frame, the center point coordinates and the direction vector of the surface area that matches the visible surface can thereby be determined. For a physical globe, the center point coordinate of the surface area is the coordinates of the center of the sphere in the reference coordinate system whose origin is the position where the camera is located, and the direction vector can be specified as the point from the center of the sphere under the same reference frame. The unit vector to the longitude and latitude are 0° points.
在某些特定的情况下,与所述可视表面匹配的表面区域仅有一个,此时该表面区域的中心点坐标和方向矢量即为所述目标实体可视表面的中心点坐标和方向矢量。而当与所述可视表面匹配的表面区域有多个时,受限于处理精度、摄像头的抖动等不可避免的因素,由视频帧获取的表面区域的中心点坐标以及方向矢量都有可能存在误差。例如,在理想情况下,同一目标实体的表面区域对应的中心点坐标应当一致,而在存在误差的情况下,多个表面区域的中心点坐标可能并不相同,并且相应的方向矢量也有所不同。若与所述可视表面匹配的两个表面区域在参考坐标系下的空间方位如图3所示,其中心点坐标分别为
Figure PCTCN2017080276-appb-000001
Figure PCTCN2017080276-appb-000002
方向矢量分别为
Figure PCTCN2017080276-appb-000003
Figure PCTCN2017080276-appb-000004
In some specific cases, there is only one surface area matching the visible surface, and the center point coordinate and the direction vector of the surface area are the center point coordinates and the direction vector of the visible surface of the target entity. . When there are a plurality of surface areas matching the visible surface, the center point coordinates and the direction vector of the surface area acquired by the video frame may exist due to factors unavoidable such as processing precision, camera shake, and the like. error. For example, in an ideal case, the coordinates of the center points corresponding to the surface areas of the same target entity should be consistent, and in the presence of errors, the center point coordinates of the multiple surface areas may not be the same, and the corresponding direction vectors may also be different. . If the spatial orientation of the two surface regions matching the visible surface in the reference coordinate system is as shown in FIG. 3, the coordinates of the center point are respectively
Figure PCTCN2017080276-appb-000001
with
Figure PCTCN2017080276-appb-000002
The direction vector is
Figure PCTCN2017080276-appb-000003
with
Figure PCTCN2017080276-appb-000004
在根据与所述可视表面匹配的表面区域的中心点坐标确定所述目标实体可视表面的中心点坐标时,若与所述可视表面匹配的表面区域有多个,将多个表面区域的中心点坐标进行加权平均处理,确定所述目标实体的中心点坐标。其中,权重的设置可以根据不同的场景具体设置,以达到最优的计算精度,而作为一种可行的实施方式,可以将所有表面区域的权重设 置为相同,此时可以根据如下公式进行计算:Determining a center point coordinate of a visible surface of the target entity according to a center point coordinate of a surface area matching the visible surface, if there are multiple surface areas matching the visible surface, a plurality of surface areas The center point coordinates are subjected to a weighted averaging process to determine the center point coordinates of the target entity. Wherein, the setting of the weights may be specifically set according to different scenarios to achieve an optimal calculation precision, and as a feasible implementation manner, the weights of all surface regions may be set. Set to the same, you can calculate according to the following formula:
Figure PCTCN2017080276-appb-000005
Figure PCTCN2017080276-appb-000005
其中,
Figure PCTCN2017080276-appb-000006
即为所述目标实体可视表面的中心点坐标,
Figure PCTCN2017080276-appb-000007
表示任意一个与所述可视表面匹配的表面区域的中心点坐标,n为匹配到的表面区域的数量。
among them,
Figure PCTCN2017080276-appb-000006
That is, the coordinates of the center point of the visible surface of the target entity,
Figure PCTCN2017080276-appb-000007
Represents the center point coordinates of any surface area that matches the visible surface, where n is the number of matched surface areas.
而在根据与所述可视表面匹配的表面区域的方向矢量确定所述目标实体可视表面方向矢量时,若与所述可视表面匹配的表面区域有多个,将多个表面区域的方向矢量加权相加并进行矢量归一化处理,获取所述目标实体可视表面的方向矢量。同理,在进行加权相加时,权重的设置可以根据不同的场景具体设置,以达到最优的计算精度,而作为一种可行的实施方式,可以将所有表面区域的权重设置为相同,此时可以根据如下公式进行计算:And when the target entity visible surface direction vector is determined according to the direction vector of the surface area matching the visible surface, if there are multiple surface areas matching the visible surface, the direction of the plurality of surface areas The vector weights are added and subjected to vector normalization processing to obtain a direction vector of the visible surface of the target entity. Similarly, when performing weighted addition, the weight setting can be specifically set according to different scenarios to achieve optimal calculation accuracy, and as a feasible implementation manner, the weights of all surface areas can be set to be the same. It can be calculated according to the following formula:
Figure PCTCN2017080276-appb-000008
Figure PCTCN2017080276-appb-000008
其中,
Figure PCTCN2017080276-appb-000009
即为所述目标实体可视表面的中心点坐标,
Figure PCTCN2017080276-appb-000010
表示任意一个与所述可视表面匹配的表面区域的方向矢量,n为匹配到的表面区域的数量。
among them,
Figure PCTCN2017080276-appb-000009
That is, the coordinates of the center point of the visible surface of the target entity,
Figure PCTCN2017080276-appb-000010
A direction vector representing any surface area that matches the visible surface, where n is the number of matched surface areas.
通过对多帧图像的连续处理,从而动态地获取每一时刻实体地球仪的空间方位,不论是改变摄像头与实体地球仪的相对位置,还是旋转实体地球仪,均能够准确地获取实体地球仪相对于摄像头的中心点坐标以及方向矢量,并基于此完成虚拟地球仪的渲染,即可保证虚拟地球仪与实体地球仪同步转动或者移动。By continuously processing the multi-frame image, the spatial orientation of the physical globe at each moment can be dynamically obtained. Whether changing the relative position of the camera and the physical globe, or rotating the physical globe, the physical globe can be accurately acquired relative to the center of the camera. The point coordinates and the direction vector, and based on this, the virtual globe is rendered to ensure that the virtual globe and the physical globe rotate or move in synchronization.
在实际场景中,可以在获取包含目标实体的视频帧之前,对目标实体进行预处理,以构建所述目标实体的预制特征集合。通过此种方式可以针对任意形状的目标实体构建相应的预制特征集合,从而完成任意形状目标实体的实时方位跟踪以及AR实现。以地球仪的场景为例,该预处理的主要过程是:对球体表面的图像特征进行离线扫描,从而得到球体表面的一个预制特征集合,这个预制特征集合会被预先存储,进而能够与实时采集 到的实体地球仪可视表面的图像特征进行比对和匹配。具体过程包括如图4所示的步骤:In an actual scenario, the target entity may be pre-processed before the video frame containing the target entity is acquired to construct a pre-made feature set of the target entity. In this way, a corresponding set of prefabricated features can be constructed for a target entity of any shape, thereby realizing real-time azimuth tracking and AR implementation of the target entity of any shape. Taking the scene of the globe as an example, the main process of the preprocessing is to scan the image features on the surface of the sphere offline to obtain a prefabricated feature set on the surface of the sphere. The prefabricated feature set is pre-stored and can be collected in real time. The image features of the visible surface of the physical globe are compared and matched. The specific process includes the steps shown in Figure 4:
步骤S401,获取目标实体多个表面区域的图像,其中,所述多个表面区域至少覆盖所述目标实体的所有表面区域;Step S401, acquiring an image of multiple surface areas of the target entity, wherein the plurality of surface areas cover at least all surface areas of the target entity;
步骤S402,对所述多个表面区域的图像进行图像特征提取,获取所述每个表面区域的图像特征信息;Step S402, performing image feature extraction on the images of the plurality of surface regions, and acquiring image feature information of each surface region;
步骤S403,根据所述每个表面区域的图像特征信息构建所述目标实体的预制特征集合。Step S403, construct a pre-made feature set of the target entity according to the image feature information of each surface area.
在对实体地球仪进行预处理时,即在采集表面区域的图像特征信息、构建预制特征集合时,需要遵循以下原则,从而保证进行方位跟踪识别时能够具有更加稳定、高效的处理性能,具体包括:When pre-processing the physical globe, that is, when collecting the image feature information of the surface area and constructing the pre-made feature set, the following principles need to be followed to ensure more stable and efficient processing performance when performing the azimuth tracking identification, including:
1)预制特征集合由球体表面的若干个表面区域(相对于整个球体表面,该表面区域为局部曲面)构成,每个表面区域的覆盖区域最大不超过一个完整的半球面;1) The prefabricated feature set is composed of a plurality of surface regions of the surface of the sphere (relative to the entire sphere surface, the surface region being a partial curved surface), and the coverage area of each surface region is no more than one complete hemisphere;
2)所有表面区域的集合应能够覆盖完整的球面;2) The collection of all surface areas should be able to cover the complete sphere;
3)每个表面区域上的图像特征应尽量充足且均匀;3) The image features on each surface area should be as adequate and uniform as possible;
4)每两个相邻的表面区域之间可具有一定的重合面积;4) each two adjacent surface areas may have a certain overlap area;
5)相邻表面区域之间的重合面积越大,越能够提升方位跟踪的精确度;5) The larger the coincidence area between adjacent surface areas, the more accurate the orientation tracking can be improved;
6)表面区域的数目越多,计算负载越大,应根据设备的实际处理能力来决定数目。6) The greater the number of surface areas, the greater the computational load, and the number should be determined based on the actual processing power of the equipment.
进一步地,本申请实施例还提供了一种实现增强现实的方法,该方法的处理流程如图5所示,包括以下步骤:Further, the embodiment of the present application further provides a method for implementing augmented reality. The processing flow of the method is as shown in FIG. 5, and includes the following steps:
步骤S501,采用前述的方位跟踪方法获取所述目标实体的中心点坐标和方向矢量。Step S501: Acquire a center point coordinate and a direction vector of the target entity by using the foregoing azimuth tracking method.
步骤S502,根据所述方向矢量由虚拟模型中获取与所述可视表面对应的显示部分。若所述目标实体为实体地球仪,则所述虚拟模型可以是与所述实体地球仪匹配的虚拟地球仪。根据所述方向矢量,基于所述虚拟地球仪的球心可以确定与所述可视表面对应的显示部分,该显示部分即为以该方向矢量为中心所指向的半球面。 Step S502, obtaining a display portion corresponding to the visible surface from the virtual model according to the direction vector. If the target entity is a physical globe, the virtual model may be a virtual globe that matches the physical globe. According to the direction vector, a display portion corresponding to the visible surface may be determined based on a center of the virtual globe, and the display portion is a hemispherical surface pointed by the direction vector.
步骤S503,根据所述中心点坐标将所述显示部分渲染至所述视频帧中,以使所述目标实体的可视表面由所述显示部分覆盖。基于中心点的坐标可以确定显示部分需要渲染至所述视频帧中的哪个位置,仍以虚拟地球仪为例,基于中心点坐标以及该虚拟地球仪的半径可以确定显示部分的半球面在该视频帧中应当处于的空间位置,由此完成最终画面的合成,使得视频帧中原本实体地球仪的可视表面由所述虚拟地球仪的相应半球面覆盖。Step S503, rendering the display portion into the video frame according to the center point coordinate, so that a visible surface of the target entity is covered by the display portion. Based on the coordinates of the center point, it can be determined at which position in the video frame the display portion needs to be rendered. Still taking the virtual globe as an example, the hemispherical surface of the display portion can be determined in the video frame based on the coordinates of the center point and the radius of the virtual globe. The spatial position that should be in, thereby completing the synthesis of the final picture, such that the visible surface of the original physical globe in the video frame is covered by the corresponding hemisphere of the virtual globe.
步骤S504,输出完成渲染的视频帧。通过前述的处理,最终呈现给用户的画面中无法看到实体地球仪,而是由虚拟地球仪替代了实体地球仪原本的位置,由于虚拟地球仪能够显示的信息量大,并且在操控实体地球仪的过程中,虚拟地球仪会以相同的角度和速度转动,能够极大地改善用户体验。Step S504, outputting the completed video frame. Through the foregoing processing, the physical globe can not be seen in the picture presented to the user, but the virtual globe replaces the original position of the physical globe, because the virtual globe can display a large amount of information, and in the process of manipulating the physical globe, The virtual globe rotates at the same angle and speed, greatly improving the user experience.
此外,除了一些普通地球仪上均具备的地理信息之外,利用虚拟地球仪可以承载较大信息量的特点,通过在虚拟模型上预先增加信息叠加点的方式,增加不同主题的虚拟信息,例如在对应的地理位置上呈现一些地标建筑、珍稀动物、地质奇观等的三维模型。具体处理方法为:In addition, in addition to the geographic information available on some common globes, virtual globes can be used to carry large amounts of information. By adding information superposition points on the virtual model in advance, virtual information of different topics can be added, for example, in correspondence. The geographical location presents three-dimensional models of landmark buildings, rare animals, and geological wonders. The specific treatment method is:
首先,获取所述显示部分中的信息叠加点。实际处理过程中,可以在获取到相应的显示部分之后,检测该显示部分中是否包含预先设置好的信息叠加点,例如在四川所在的位置预先设置一个关于大熊猫的信息叠加点,当虚拟地球仪显示部分的半球面包含四川所在的地理区域时,即可以检测到该信息叠加点。First, an information superposition point in the display portion is acquired. In the actual processing, after the corresponding display part is obtained, it is detected whether the display part contains a preset information superposition point, for example, an information superposition point about the giant panda is preset in the position of Sichuan, when the virtual globe is set. When the hemispherical surface of the display part contains the geographical area where Sichuan is located, the superposition point of the information can be detected.
然后,将所述信息叠加点对应的虚拟信息渲染至所述视频帧中,以使所述显示部分的相应位置显示所述虚拟信息。所述虚拟信息可以是一个大熊猫的三维模型,由此用户可以直接在屏幕上看到一个大熊猫位于四川的位置,更加直观和清晰地了解各类地理知识。Then, the virtual information corresponding to the information superposition point is rendered into the video frame, so that the corresponding position of the display portion displays the virtual information. The virtual information can be a three-dimensional model of a giant panda, so that the user can directly see a giant panda in Sichuan on the screen, and more intuitive and clear understanding of various geographical knowledge.
进一步地,为了能够向用户传递更形象、更动态的信息,可以在虚拟模型上设置一些触发点。在渲染过程中,判断所述显示部分的触发点是否位于所述视频帧的触发区域;若所述触发点处于所述视频帧的触发区域,将所述触发点的触发效果渲染至所述视频帧中。Further, in order to be able to deliver more avatar and more dynamic information to the user, some trigger points can be set on the virtual model. During a rendering process, determining whether a trigger point of the display portion is located in a trigger area of the video frame; if the trigger point is in a trigger area of the video frame, rendering a trigger effect of the trigger point to the video In the frame.
其中,所述触发区域可以根据实际需求设置。例如对于触屏,可以将 该触发区域设置于用户当前触摸的区域,以实现用户点击实现触发;也可以静态地设置在屏幕的某一块特定区域,例如屏幕中央;还可以进一步通过光标、光圈等进行标记,以便于用户识别。而所述触发效果可以是语音、视频、三维动画效果等,例如将所述触发点设置于大熊猫的显示位置,当该大熊猫的三维模型处于触发区域内时,大熊猫的三维模型会被放大,同时进行相应的动作,例如走路、起立等;此外,触发效果也可以是突出显示效果,例如将触发点设置为任意行政区域的范围,当某一行政区域进入到触发区域时,该行政区域将会高亮显示。在此,本领域技术人员应当理解,所述触发区域、触发效果以及触发点的设置方式仅为举例,其他现有的或者今后可能出现的方式如可适用于本申请,也应包含在本申请保护范围以内,并在此以引用的方式包含于此。The trigger area can be set according to actual needs. For example, for a touch screen, you can The trigger area is set in the area currently touched by the user, so that the user clicks to implement the trigger; or can be statically set in a specific area of the screen, such as the center of the screen; and can be further marked by a cursor, an aperture, etc., to facilitate user identification. . The triggering effect may be a voice, a video, a three-dimensional animation effect, etc., for example, setting the trigger point to a display position of the giant panda. When the three-dimensional model of the giant panda is in the triggering region, the three-dimensional model of the giant panda is Zoom in and perform corresponding actions at the same time, such as walking, standing, etc. In addition, the triggering effect can also be a highlighting effect, for example, setting the trigger point to the range of any administrative area, when an administrative area enters the triggering area, the administrative The area will be highlighted. Herein, those skilled in the art should understand that the triggering area, the triggering effect, and the setting manner of the triggering point are only examples, and other existing or future possible manners may be applied to the present application, and should also be included in the present application. It is within the scope of protection and is hereby incorporated by reference.
基于本申请的另一方面,还提供了一种目标实体的方位跟踪装置60,该装置60用于在实现AR的处理过程中,实现对实体目标可视表面的方位跟踪,其具体结构如图6所示,包括图像采集模块610、特征提取模块620、特征匹配模块630和综合处理模块640。具体地,所述图像采集模块610用于获取包含目标实体的视频帧。所述目标实体可以是实体地球仪,也可以是例如球体、椭球体、多面体等其它规则或者不规则形状的实体。在实际场景中可以通过带有摄像头等图像采集装置的设备来拍摄目标实体,以获取到包含目标实体的视频帧。例如,通过智能手机的摄像头拍摄一段关于实体地球仪的视频,该视频中的任意一帧均可用于后续的处理,以获取在该视频帧中实体地球仪可视表面的中心坐标点和方向矢量。Based on another aspect of the present application, an orientation tracking device 60 for a target entity is also provided. The device 60 is configured to implement azimuth tracking of a visible surface of a physical target during processing of implementing the AR. 6 shows an image acquisition module 610, a feature extraction module 620, a feature matching module 630, and an integrated processing module 640. Specifically, the image collection module 610 is configured to acquire a video frame that includes a target entity. The target entity may be a physical globe, or may be other regular or irregularly shaped entities such as spheres, ellipsoids, polyhedra, and the like. In an actual scenario, a target entity can be photographed by a device with an image capture device such as a camera to obtain a video frame containing the target entity. For example, a video of a physical globe is taken by a camera of a smartphone, and any frame in the video can be used for subsequent processing to obtain a central coordinate point and a direction vector of the visible surface of the physical globe in the video frame.
所述特征提取模块620用于对所述视频帧中的目标实体进行图像特征提取,获取所述目标实体可视表面的图像特征信息。对于某一目标实体,在视频帧中仅可以看到其某一部分的表面,例如对于球形的实体地球仪,视频帧中该实体地球仪的可视表面最大为半个球面,在进行图像特征提取时,也仅需要提取可视的半个球面的图像特征信息。而对于例如多面体或者其它不规则形状的实体,可视表面根据摄像头镜头角度的不同,可能包含其中的一个或者多个侧面。在进行图像特征提取时,可以采用现有技术中各类成熟的图像特征提取技术,例如基于颜色特征、纹理特征、形状特 征、空间关系特征等图像特征信息中任意一种或多种的组合,提取地球仪表面图案的一种或多种图像特征信息。The feature extraction module 620 is configured to perform image feature extraction on a target entity in the video frame to acquire image feature information of a visible surface of the target entity. For a certain target entity, only the surface of a certain part of the surface can be seen in the video frame. For example, for a spherical physical globe, the visible surface of the physical globe in the video frame is at most half a spherical surface, and when image feature extraction is performed, It is also only necessary to extract visual half-spherical image feature information. For entities such as polyhedrons or other irregular shapes, the visible surface may include one or more of the sides depending on the angle of the camera lens. In the image feature extraction, various mature image feature extraction techniques in the prior art can be used, for example, based on color features, texture features, and shapes. A combination of any one or more of image feature information such as signs and spatial relationship features, and extracting one or more image feature information of the surface pattern of the globe.
所述特征匹配模块630用于基于所述目标实体的预制特征集合对获取到的所述目标实体可视表面的图像特征信息进行特征识别,在预制特征集合中获取与所述可视表面匹配的表面区域。其中,所述预制特征集合包含所述目标实体的多个表面区域以及每个表面区域的图像特征信息,仍以前述实体地球仪为例,其预制特征集合中包含的多个表面区域的一种可行的划分方式为:以地球仪的经线方向进行划分,将整个地球仪划分为4个完整的半球面,这4个半球面的中心点(即球心)完全重合,且相邻的每两个半球面之间具有90°的重合区域,这种划分方式在赤道平面上的投影如图2(a)所示,其中各个半球面在赤道平面上的投影如图2(b)~(e)。所述预制特征集合可以直接采用已有的集合数据,也可以在进行处理之前,通过离线预处理的方式获取。The feature matching module 630 is configured to perform feature recognition on the acquired image feature information of the target entity visible surface based on the set of pre-made features of the target entity, and acquire a matching match with the visible surface in the pre-made feature set. Surface area. Wherein the prefabricated feature set includes a plurality of surface regions of the target entity and image feature information of each surface region, and the physical globe is still taken as an example, and a plurality of surface regions included in the prefabricated feature set are feasible The division is as follows: dividing the globe by the direction of the warp, dividing the whole globe into four complete hemispheres, the center points of the four hemispheres (ie, the center of the sphere) are completely coincident, and each adjacent two hemispheres There is a 90° overlap between the two. The projection of this division on the equatorial plane is shown in Fig. 2(a), and the projection of each hemisphere on the equatorial plane is shown in Fig. 2(b) to (e). The preset feature set may directly adopt the existing set data, or may be obtained by offline pre-processing before processing.
而由于每个半球面上的地理环境不同,相应部分的图案也会各不相同,例如水域和陆地会采用不同的色彩,陆地中各种类型的地形同样会采用不同的颜色或图案等,由此4个半球面对应的图像特征信息也将完全不同。所述预制特征集合中,每个半球面对应的图像特征信息同样可以采用颜色特征、纹理特征、形状特征、空间关系特征等图像特征信息中任意一种或多种的组合。通过图像特征的匹配识别,若可视表面某一部分的图像特征信息与预制特征集合中的某一表面区域中的相应部分相同或者相似度超过阈值,则可以确定两者匹配,由此获取预制特征集合中所有与可视表面匹配的表面区域。Because the geographical environment on each hemisphere is different, the patterns of the corresponding parts will be different. For example, the water and land will adopt different colors, and various types of terrain on the land will also adopt different colors or patterns. The image feature information corresponding to the four hemispheres will also be completely different. In the prefabricated feature set, the image feature information corresponding to each hemispherical surface may also adopt a combination of any one or more of image feature information such as a color feature, a texture feature, a shape feature, and a spatial relationship feature. By matching the image features, if the image feature information of a certain part of the visible surface is the same as the corresponding part of a certain surface area in the prefabricated feature set or the similarity exceeds the threshold, the two can be matched, thereby obtaining the prefabricated feature. All surface areas in the collection that match the visible surface.
在实际场景中,由于在视频获取过程中摄像头与实体地球仪的相对位置的不同,获取到的实体地球仪的可视表面可能包含了上述预制特征集合中一个完整的半球面,也可能由其中多个半球面的某一部分组成。例如,镜头方向与赤道上任意一点和地心连线重合,此时实体地球仪的可视表面可能包含上述预制特征集合中一个完整的半球面,或者由其中两个半球面的一部分组成;若镜头方向与地轴重合,则此时实体地球仪的可视表面可能由上述预制特征集合中4个半球面的各一部分组成。 In an actual scenario, due to the difference in the relative positions of the camera and the physical globe during the video acquisition process, the visible surface of the acquired physical globe may contain a complete hemisphere in the set of prefabricated features, or may be composed of multiple A part of the hemisphere. For example, the lens direction coincides with any point on the equator and the geocentric line, and the visible surface of the physical globe may contain a complete hemisphere in the set of prefabricated features, or a part of two hemispheres; The direction coincides with the earth axis, at which point the visible surface of the physical globe may consist of portions of the four hemispheres in the set of prefabricated features described above.
所述综合处理模块640用于根据与所述可视表面匹配的表面区域的中心点坐标和方向矢量分别确定所述目标实体可视表面的中心点坐标和方向矢量。本实施例在进行数学处理的过程中,将当前摄像头所在的位置设置为参考坐标系的原点,在此参考坐标系下,不论是摄像头移动还是目标实体移动,都可以视为目标实体的位置发生了变化,而摄像头的位置保持不变。由于目标实体可视表面的空间位置可以在视频帧中确定,则由此可以确定与所述可视表面匹配的表面区域的中心点坐标和方向矢量。对于实体地球仪,其表面区域的所述中心点坐标即为以摄像头所在的位置为原点的参考坐标系下球心的坐标,而所述方向矢量则可指定为在同样参考系下从球心点到经度、纬度均为0°点的单位矢量。The comprehensive processing module 640 is configured to determine a center point coordinate and a direction vector of the visible surface of the target entity according to a center point coordinate and a direction vector of the surface area matched with the visible surface. In the process of performing mathematical processing, the position of the current camera is set as the origin of the reference coordinate system. In this reference coordinate system, whether the camera moves or the target entity moves, the position of the target entity can be regarded as occurring. The change, while the position of the camera remains the same. Since the spatial position of the visible surface of the target entity can be determined in the video frame, the center point coordinates and the direction vector of the surface area that matches the visible surface can thereby be determined. For a physical globe, the center point coordinate of the surface area is the coordinates of the center of the sphere in the reference coordinate system whose origin is the position where the camera is located, and the direction vector can be specified as the point from the center of the sphere under the same reference frame. The unit vector to the longitude and latitude are 0° points.
在某些特定的情况下,与所述可视表面匹配的表面区域仅有一个,此时该表面区域的中心点坐标和方向矢量即为所述目标实体可视表面的中心点坐标和方向矢量。而当与所述可视表面匹配的表面区域有多个时,受限于处理精度、摄像头的抖动等不可避免的因素,由视频帧获取的表面区域的中心点坐标以及方向矢量都有可能存在误差。例如,在理想情况下,同一目标实体的表面区域对应的中心点坐标应当一致,而在存在误差的情况下,多个表面区域的中心点坐标可能并不相同,并且相应的方向矢量也有所不同。若与所述可视表面匹配的两个表面区域在参考坐标系下的空间方位如图3所示,其中心点坐标分别为
Figure PCTCN2017080276-appb-000011
Figure PCTCN2017080276-appb-000012
方向矢量分别为
Figure PCTCN2017080276-appb-000013
Figure PCTCN2017080276-appb-000014
In some specific cases, there is only one surface area matching the visible surface, and the center point coordinate and the direction vector of the surface area are the center point coordinates and the direction vector of the visible surface of the target entity. . When there are a plurality of surface areas matching the visible surface, the center point coordinates and the direction vector of the surface area acquired by the video frame may exist due to factors unavoidable such as processing precision, camera shake, and the like. error. For example, in an ideal case, the coordinates of the center points corresponding to the surface areas of the same target entity should be consistent, and in the presence of errors, the center point coordinates of the multiple surface areas may not be the same, and the corresponding direction vectors may also be different. . If the spatial orientation of the two surface regions matching the visible surface in the reference coordinate system is as shown in FIG. 3, the coordinates of the center point are respectively
Figure PCTCN2017080276-appb-000011
with
Figure PCTCN2017080276-appb-000012
The direction vector is
Figure PCTCN2017080276-appb-000013
with
Figure PCTCN2017080276-appb-000014
所述综合处理模块640在根据与所述可视表面匹配的表面区域的中心点坐标确定所述目标实体可视表面的中心点坐标时,若与所述可视表面匹配的表面区域有多个,将多个表面区域的中心点坐标进行加权平均处理,确定所述目标实体的中心点坐标。其中,权重的设置可以根据不同的场景具体设置,以达到最优的计算精度,而作为一种可行的实施方式,可以将所有表面区域的权重设置为相同,此时可以根据如下公式进行计算:When the integrated processing module 640 determines the center point coordinates of the visible surface of the target entity according to the center point coordinates of the surface area matching the visible surface, if there are multiple surface areas matching the visible surface And performing weighted averaging processing on center point coordinates of the plurality of surface regions to determine a center point coordinate of the target entity. The weight setting may be specifically set according to different scenarios to achieve an optimal calculation precision, and as a feasible implementation manner, the weights of all surface regions may be set to be the same, and the calculation may be performed according to the following formula:
Figure PCTCN2017080276-appb-000015
Figure PCTCN2017080276-appb-000015
其中,
Figure PCTCN2017080276-appb-000016
即为所述目标实体可视表面的中心点坐标,
Figure PCTCN2017080276-appb-000017
表示任意一个 与所述可视表面匹配的表面区域的中心点坐标,n为匹配到的表面区域的数量。
among them,
Figure PCTCN2017080276-appb-000016
That is, the coordinates of the center point of the visible surface of the target entity,
Figure PCTCN2017080276-appb-000017
Represents the coordinates of the center point of any surface area that matches the visible surface, where n is the number of matched surface areas.
而所述综合处理模块640在根据与所述可视表面匹配的表面区域的方向矢量确定所述目标实体可视表面方向矢量时,若与所述可视表面匹配的表面区域有多个,将多个表面区域的方向矢量加权相加并进行矢量归一化处理,获取所述目标实体可视表面的方向矢量。同理,在进行加权相加时,权重的设置可以根据不同的场景具体设置,以达到最优的计算精度,而作为一种可行的实施方式,可以将所有表面区域的权重设置为相同,此时可以根据如下公式进行计算:When the integrated processing module 640 determines the target entity visible surface direction vector according to the direction vector of the surface area matching the visible surface, if there are multiple surface areas matching the visible surface, The direction vectors of the plurality of surface regions are weighted and added and subjected to vector normalization processing to obtain a direction vector of the visible surface of the target entity. Similarly, when performing weighted addition, the weight setting can be specifically set according to different scenarios to achieve optimal calculation accuracy, and as a feasible implementation manner, the weights of all surface areas can be set to be the same. It can be calculated according to the following formula:
Figure PCTCN2017080276-appb-000018
Figure PCTCN2017080276-appb-000018
其中,
Figure PCTCN2017080276-appb-000019
即为所述目标实体可视表面的中心点坐标,
Figure PCTCN2017080276-appb-000020
表示任意一个与所述可视表面匹配的表面区域的方向矢量,n为匹配到的表面区域的数量。
among them,
Figure PCTCN2017080276-appb-000019
That is, the coordinates of the center point of the visible surface of the target entity,
Figure PCTCN2017080276-appb-000020
A direction vector representing any surface area that matches the visible surface, where n is the number of matched surface areas.
通过对多帧图像的连续处理,从而动态地获取每一时刻实体地球仪的空间方位,不论是改变摄像头与实体地球仪的相对位置,还是旋转实体地球仪,均能够准确地获取实体地球仪相对于摄像头的中心点坐标以及方向矢量,并基于此完成虚拟地球仪的渲染,即可保证虚拟地球仪与实体地球仪同步转动或者移动。By continuously processing the multi-frame image, the spatial orientation of the physical globe at each moment can be dynamically obtained. Whether changing the relative position of the camera and the physical globe, or rotating the physical globe, the physical globe can be accurately acquired relative to the center of the camera. The point coordinates and the direction vector, and based on this, the virtual globe is rendered to ensure that the virtual globe and the physical globe rotate or move in synchronization.
在实际场景中,可以通过预处理模块在获取包含目标实体的视频帧之前,对目标实体进行预处理,以构建所述目标实体的预制特征集合。通过此种方式可以针对任意形状的目标实体构建相应的预制特征集合,从而完成任意形状目标实体的实时方位跟踪以及AR实现。以地球仪的场景为例,该预处理的主要过程是:对球体表面的图像特征进行离线扫描,从而得到球体表面的一个预制特征集合,这个预制特征集合会被预先存储,进而能够与实时采集到的实体地球仪可视表面的图像特征进行比对和匹配。所述预处理模块具体用于控制所述图像采集模块获取目标实体多个表面区域的图像,其中,所述多个表面区域至少覆盖所述目标实体的所有表面区域;控制所述特征提取模块对所述多个表面区域的图像进行图像特征提取,获 取所述每个表面区域的图像特征信息;以及根据所述每个表面区域的图像特征信息构建所述目标实体的预制特征集合。In an actual scenario, the target entity may be pre-processed by the pre-processing module to obtain a pre-made feature set of the target entity before acquiring the video frame including the target entity. In this way, a corresponding set of prefabricated features can be constructed for a target entity of any shape, thereby realizing real-time azimuth tracking and AR implementation of the target entity of any shape. Taking the scene of the globe as an example, the main process of the preprocessing is to scan the image features on the surface of the sphere offline to obtain a prefabricated feature set on the surface of the sphere. The prefabricated feature set is pre-stored and can be collected in real time. The physical features of the physical globe are aligned and matched to the image features of the visible surface. The pre-processing module is specifically configured to control the image acquisition module to acquire an image of a plurality of surface regions of the target entity, wherein the plurality of surface regions cover at least all surface regions of the target entity; and the feature extraction module is controlled Image of the plurality of surface regions is subjected to image feature extraction Taking image feature information of each surface area; and constructing a set of prefabricated features of the target entity according to image feature information of each surface area.
在对实体地球仪进行预处理时,即在采集表面区域的图像特征信息、构建预制特征集合时,需要遵循以下原则,从而保证进行方位跟踪识别时能够具有更加稳定、高效的处理性能,具体包括:When pre-processing the physical globe, that is, when collecting the image feature information of the surface area and constructing the pre-made feature set, the following principles need to be followed to ensure more stable and efficient processing performance when performing the azimuth tracking identification, including:
1)预制特征集合由球体表面的若干个表面区域(相对于整个球体表面,该表面区域为局部曲面)构成,每个表面区域的覆盖区域最大不超过一个完整的半球面;1) The prefabricated feature set is composed of a plurality of surface regions of the surface of the sphere (relative to the entire sphere surface, the surface region being a partial curved surface), and the coverage area of each surface region is no more than one complete hemisphere;
2)所有表面区域的集合应能够覆盖完整的球面;2) The collection of all surface areas should be able to cover the complete sphere;
3)每个表面区域上的图像特征应尽量充足且均匀;3) The image features on each surface area should be as adequate and uniform as possible;
4)每两个相邻的表面区域之间可具有一定的重合面积;4) each two adjacent surface areas may have a certain overlap area;
5)相邻表面区域之间的重合面积越大,越能够提升方位跟踪的精确度;5) The larger the coincidence area between adjacent surface areas, the more accurate the orientation tracking can be improved;
6)表面区域的数目越多,计算负载越大,应根据设备的实际处理能力来决定数目。6) The greater the number of surface areas, the greater the computational load, and the number should be determined based on the actual processing power of the equipment.
进一步地,本申请实施例还提供了一种实现增强现实的设备,该设备的结构如图7所示,包括方位跟踪装置60、渲染装置710和输出装置720。具体地,所述方位跟踪装置60用于获取所述目标实体的中心点坐标和方向矢量。所述渲染装置710用于根据所述方向矢量由虚拟模型中获取与所述可视表面对应的显示部分,以及根据所述中心点坐标将所述显示部分渲染至所述视频帧中,以使所述目标实体的可视表面由所述显示部分覆盖。若所述目标实体为实体地球仪,则所述虚拟模型可以是与所述实体地球仪匹配的虚拟地球仪。根据所述方向矢量,基于所述虚拟地球仪的球心可以确定与所述可视表面对应的显示部分,该显示部分即为以该方向矢量为中心所指向的半球面。基于中心点的坐标可以确定显示部分需要渲染至所述视频帧中的哪个位置,仍以虚拟地球仪为例,基于中心点坐标以及该虚拟地球仪的半径可以确定显示部分的半球面在该视频帧中应当处于的空间位置,由此完成最终画面的合成,使得视频帧中原本实体地球仪的可视表面由所述虚拟地球仪的相应半球面覆盖。Further, the embodiment of the present application further provides an apparatus for implementing augmented reality. The structure of the apparatus is as shown in FIG. 7, and includes an azimuth tracking device 60, a rendering device 710, and an output device 720. Specifically, the azimuth tracking device 60 is configured to acquire a center point coordinate and a direction vector of the target entity. The rendering device 710 is configured to acquire a display portion corresponding to the visible surface from a virtual model according to the direction vector, and render the display portion into the video frame according to the center point coordinate, so that The visible surface of the target entity is covered by the display portion. If the target entity is a physical globe, the virtual model may be a virtual globe that matches the physical globe. According to the direction vector, a display portion corresponding to the visible surface may be determined based on a center of the virtual globe, and the display portion is a hemispherical surface pointed by the direction vector. Based on the coordinates of the center point, it can be determined at which position in the video frame the display portion needs to be rendered. Still taking the virtual globe as an example, the hemispherical surface of the display portion can be determined in the video frame based on the coordinates of the center point and the radius of the virtual globe. The spatial position that should be in, thereby completing the synthesis of the final picture, such that the visible surface of the original physical globe in the video frame is covered by the corresponding hemisphere of the virtual globe.
所述输出装置720用于输出完成渲染的视频帧。通过前述的处理,最 终呈现给用户的画面中无法看到实体地球仪,而是由虚拟地球仪替代了实体地球仪原本的位置,由于虚拟地球仪能够显示的信息量大,并且在操控实体地球仪的过程中,虚拟地球仪会以相同的角度和速度转动,能够极大地改善用户体验。The output device 720 is configured to output a video frame that is finished rendering. Through the aforementioned processing, the most The physical globe is not visible in the picture presented to the user, but the virtual globe replaces the original position of the physical globe. The virtual globe can display the same amount of information, and the virtual globe will be the same during the manipulation of the physical globe. The angle and speed of rotation can greatly improve the user experience.
在此,本领域技术人员应当理解,所述设备可以包括但不限于用户终端或者用户终端与网络设备通过网络相集成所构成的设备。所述用户终端包括但不限于个人计算机、触控终端等实现,具体可以是智能手机、平板电脑、PDA、AR眼镜或者其它具备图像获取、处理以及输出功能的电子设备;所述网络设备包括但不限于如网络主机、单个网络服务器、多个网络服务器集或基于云计算的计算机集合等实现。在此,云由基于云计算(Cloud Computing)的大量主机或网络服务器构成,其中,云计算是分布式计算的一种,由一群松散耦合的计算机集组成的一个虚拟计算机。此外,所述设备还可以是运行有包含相关算法及图形用户界面的应用程序(APP)的电子设备,或者应用程序本身。Here, those skilled in the art should understand that the device may include, but is not limited to, a user terminal or a device formed by integrating a user terminal and a network device through a network. The user terminal includes, but is not limited to, a personal computer, a touch terminal, and the like, and specifically may be a smart phone, a tablet computer, a PDA, an AR glasses, or other electronic device having an image acquisition, processing, and output function; Implementations are not limited to, for example, a network host, a single network server, multiple network server sets, or a cloud-based computer collection. Here, the cloud is composed of a large number of host or network servers based on Cloud Computing, which is a kind of distributed computing, a virtual computer composed of a group of loosely coupled computers. Furthermore, the device may also be an electronic device running an application (APP) containing associated algorithms and graphical user interfaces, or the application itself.
以智能手机为例,作为一种典型的实现方式,智能手机的摄像头可用于实现图像采集模块610的相关功能,其处理器可用于实现特征提取模块620、特征匹配模块630、综合处理模块640以及渲染装置710的相关功能,而其屏幕则可实现输出装置720的相关功能。在实际场景下,处理器也可以通过通信模块将相关数据上传网络设备,由网络设备完成相关数据的计算、处理工作。Taking a smart phone as an example, as a typical implementation, the camera of the smart phone can be used to implement the related functions of the image capturing module 610, and the processor can be used to implement the feature extraction module 620, the feature matching module 630, the comprehensive processing module 640, and The related functions of the device 710 are rendered while their screens enable the associated functions of the output device 720. In an actual scenario, the processor can also upload relevant data to the network device through the communication module, and the network device completes calculation and processing of related data.
此外,除了一些普通地球仪上均具备的地理信息之外,利用虚拟地球仪可以承载较大信息量的特点,通过在虚拟模型上预先增加信息叠加点的方式,增加不同主题的虚拟信息,例如在对应的地理位置上呈现一些地标建筑、珍稀动物、地质奇观等的三维模型。由此,所述渲染装置710还用于进行如下处理:In addition, in addition to the geographic information available on some common globes, virtual globes can be used to carry large amounts of information. By adding information superposition points on the virtual model in advance, virtual information of different topics can be added, for example, in correspondence. The geographical location presents three-dimensional models of landmark buildings, rare animals, and geological wonders. Thus, the rendering device 710 is further configured to perform the following processing:
首先,获取所述显示部分中的信息叠加点。实际处理过程中,可以在获取到相应的显示部分之后,检测该显示部分中是否包含预先设置好的信息叠加点,例如在四川所在的位置预先设置一关于大熊猫的信息叠加点,当虚拟地球仪显示部分的半球面包含四川所在的地理区域时,即可以检测 到该信息叠加点。First, an information superposition point in the display portion is acquired. In the actual processing, after the corresponding display part is obtained, it is detected whether the display part contains a pre-set information superposition point, for example, a superposition point about the giant panda is preset in the position of Sichuan, when the virtual globe is set. When the hemisphere of the display part contains the geographical area where Sichuan is located, it can be detected. Go to the superposition point of this information.
然后,将所述信息叠加点对应的虚拟信息渲染至所述视频帧中,以使所述显示部分的相应位置显示所述虚拟信息。所述虚拟信息可以是一个大熊猫的三维模型,由此用户可以直接在屏幕上看到一个大熊猫位于四川的位置,更加直观和清晰地了解各类地理知识。Then, the virtual information corresponding to the information superposition point is rendered into the video frame, so that the corresponding position of the display portion displays the virtual information. The virtual information can be a three-dimensional model of a giant panda, so that the user can directly see a giant panda in Sichuan on the screen, and more intuitive and clear understanding of various geographical knowledge.
进一步地,为了能够向用户传递更形象、更动态的信息,可以在虚拟模型上设置一些触发点。在渲染过程中,所述渲染装置710判断所述显示部分的触发点是否位于所述视频帧的触发区域;若所述触发点处于所述视频帧的触发区域,将所述触发点的触发效果渲染至所述视频帧中。Further, in order to be able to deliver more avatar and more dynamic information to the user, some trigger points can be set on the virtual model. During the rendering process, the rendering device 710 determines whether the trigger point of the display portion is located in a trigger region of the video frame; if the trigger point is in a trigger region of the video frame, triggering the trigger point Rendered into the video frame.
其中,所述触发区域可以根据实际需求设置。例如对于触屏,可以将该触发区域设置于用户当前触摸的区域,以实现用户点击实现触发;也可以静态地设置在屏幕的某一块特定区域,例如屏幕中央;还可以进一步通过光标、光圈等进行标记,以便于用户识别。而所述触发效果可以是音频、视频、三维动画效果等,例如将所述触发点设置于大熊猫的显示位置,当该大熊猫的三维模型处于触发区域内时,大熊猫的三维模型会被放大,同时进行相应的动作,例如走路、起立等;此外,触发效果也可以是突出显示效果,例如将触发点设置为任意行政区域的范围,当某一行政区域进入到触发区域时,该行政区域将会高亮显示。在此,本领域技术人员应当理解,所述触发区域、触发效果以及触发点的设置方式仅为举例,其他现有的或者今后可能出现的方式如可适用于本申请,也应包含在本申请保护范围以内,并在此以引用的方式包含于此。The trigger area can be set according to actual needs. For example, for a touch screen, the trigger area may be set to an area currently touched by the user to implement a user click to implement triggering; or may be statically set in a certain specific area of the screen, such as the center of the screen; and may further pass a cursor, an aperture, etc. Mark it for user identification. The triggering effect may be audio, video, three-dimensional animation effect, etc., for example, setting the trigger point to the display position of the giant panda. When the three-dimensional model of the giant panda is in the triggering area, the three-dimensional model of the giant panda will be Zoom in and perform corresponding actions at the same time, such as walking, standing, etc. In addition, the triggering effect can also be a highlighting effect, for example, setting the trigger point to the range of any administrative area, when an administrative area enters the triggering area, the administrative The area will be highlighted. Herein, those skilled in the art should understand that the triggering area, the triggering effect, and the setting manner of the triggering point are only examples, and other existing or future possible manners may be applied to the present application, and should also be included in the present application. It is within the scope of protection and is hereby incorporated by reference.
综上所述,本申请的方案能够基于增强现实技术将一些虚拟的图形、图像、信息等叠加到真实场景中去,起到了对现实场景进行增强的作用,通过将实体地球仪作为目标实体,使其能够同时兼备现有技术中的实体地球仪和虚拟地球仪的优点。通过实时获取目标实体的中心坐标点和方向矢量实现方位跟踪,并基于该中心坐标点和方向矢量对输出的视频帧进行实时渲染,使得用户在操控实体地球仪的过程中,虚拟地球仪会以相同的角度和速度转动,能够极大地改善用户体验。In summary, the solution of the present application can superimpose some virtual graphics, images, information, etc. into a real scene based on augmented reality technology, and enhances the real scene by using the physical globe as a target entity. It combines the advantages of both physical globes and virtual globes in the prior art. Azimuth tracking is realized by acquiring the central coordinate point and direction vector of the target entity in real time, and the output video frame is rendered in real time based on the central coordinate point and the direction vector, so that the virtual globe will be the same in the process of manipulating the physical globe. Angle and speed rotation can greatly improve the user experience.
此外,通过实时获取的包含目标实体的视频帧,进行图像特征提取, 并与预先构建的预制特征集合进行图像特征匹配,并基于匹配结果进行计算,获取当前视频帧中目标实体可视表面的中心点坐标和方向矢量,由此能够快速准确地实现目标实体的方位跟踪,保证屏幕中显示的虚拟地球仪与实体地球仪能够同步运动,给予用户真实的空间存在感和实际的操作体验感。In addition, image feature extraction is performed by a video frame containing the target entity acquired in real time, Image feature matching is performed with the pre-built pre-made feature set, and calculation is performed based on the matching result to obtain the center point coordinate and the direction vector of the visible surface of the target entity in the current video frame, thereby enabling fast and accurate orientation tracking of the target entity. To ensure that the virtual globe displayed on the screen and the physical globe can move synchronously, giving the user a real sense of space and actual operational experience.
需要注意的是,本申请可在软件和/或软件与硬件的组合体中被实施,例如,可采用专用集成电路(ASIC)、通用目的计算机或任何其他类似硬件设备来实现。在一个实施例中,本申请的软件程序可以通过处理器执行以实现上文所述步骤或功能。同样地,本申请的软件程序(包括相关的数据结构)可以被存储到计算机可读记录介质中,例如,RAM存储器,磁或光驱动器或软磁盘及类似设备。另外,本申请的一些步骤或功能可采用硬件来实现,例如,作为与处理器配合从而执行各个步骤或功能的电路。It should be noted that the present application can be implemented in software and/or a combination of software and hardware, for example, using an application specific integrated circuit (ASIC), a general purpose computer, or any other similar hardware device. In one embodiment, the software program of the present application can be executed by a processor to implement the steps or functions described above. Likewise, the software programs (including related data structures) of the present application can be stored in a computer readable recording medium such as a RAM memory, a magnetic or optical drive or a floppy disk and the like. In addition, some of the steps or functions of the present application may be implemented in hardware, for example, as a circuit that cooperates with a processor to perform various steps or functions.
另外,本申请的一部分可被应用为计算机程序产品,例如计算机程序指令,当其被计算机执行时,通过该计算机的操作,可以调用或提供根据本申请的方法和/或技术方案。而调用本申请的方法的程序指令,可能被存储在固定的或可移动的记录介质中,和/或通过广播或其他信号承载媒体中的数据流而被传输,和/或被存储在根据所述程序指令运行的计算机设备的工作存储器中。在此,根据本申请的一个实施例包括一个装置,该装置包括用于存储计算机程序指令的存储器和用于执行程序指令的处理器,其中,当该计算机程序指令被该处理器执行时,触发该装置运行基于前述根据本申请的多个实施例的方法和/或技术方案。In addition, a portion of the present application can be applied as a computer program product, such as computer program instructions, which, when executed by a computer, can invoke or provide a method and/or technical solution in accordance with the present application. The program instructions for invoking the method of the present application may be stored in a fixed or removable recording medium, and/or transmitted by a data stream in a broadcast or other signal bearing medium, and/or stored in a The working memory of the computer device in which the program instructions are run. Herein, an embodiment in accordance with the present application includes a device including a memory for storing computer program instructions and a processor for executing program instructions, wherein when the computer program instructions are executed by the processor, triggering The apparatus operates based on the aforementioned methods and/or technical solutions in accordance with various embodiments of the present application.
对于本领域技术人员而言,显然本申请不限于上述示范性实施例的细节,而且在不背离本申请的精神或基本特征的情况下,能够以其他的具体形式实现本申请。因此,无论从哪一点来看,均应将实施例看作是示范性的,而且是非限制性的,本申请的范围由所附权利要求而不是上述说明限定,因此旨在将落在权利要求的等同要件的含义和范围内的所有变化涵括在本申请内。不应将权利要求中的任何附图标记视为限制所涉及的权利要求。此外,显然“包括”一词不排除其他单元或步骤,单数不排除复数。装置权利要求中陈述的多个单元或装置也可以由一个单元或装置通过软件 或者硬件来实现。第一、第二等词语用来表示名称,而并不表示任何特定的顺序。 It is obvious to those skilled in the art that the present application is not limited to the details of the above-described exemplary embodiments, and the present invention can be implemented in other specific forms without departing from the spirit or essential characteristics of the present application. Therefore, the present embodiments are to be considered as illustrative and not restrictive, and the scope of the invention is defined by the appended claims instead All changes in the meaning and scope of equivalent elements are included in this application. Any reference signs in the claims should not be construed as limiting the claim. In addition, it is to be understood that the word "comprising" does not exclude other elements or steps. A plurality of units or devices recited in the device claims may also be implemented by a unit or device. Or hardware to achieve. The words first, second, etc. are used to denote names and do not denote any particular order.

Claims (18)

  1. 一种目标实体的方位跟踪方法,其中,该方法包括:A method for tracking an orientation of a target entity, wherein the method comprises:
    获取包含目标实体的视频帧;Get a video frame containing the target entity;
    对所述视频帧中的目标实体进行图像特征提取,获取所述目标实体可视表面的图像特征信息;Performing image feature extraction on the target entity in the video frame to acquire image feature information of the visible surface of the target entity;
    基于所述目标实体的预制特征集合对获取到的所述目标实体可视表面的图像特征信息进行特征识别,在预制特征集合中获取与所述可视表面匹配的表面区域,其中,所述预制特征集合包含所述目标实体的多个表面区域以及每个表面区域的图像特征信息;Performing feature recognition on the acquired image feature information of the visible surface of the target entity based on the set of prefabricated features of the target entity, and acquiring a surface area matching the visible surface in the set of prefabricated features, wherein the prefabrication The feature set includes a plurality of surface regions of the target entity and image feature information of each surface region;
    根据与所述可视表面匹配的表面区域的中心点坐标和方向矢量分别确定所述目标实体可视表面的中心点坐标和方向矢量。A center point coordinate and a direction vector of the visible surface of the target entity are respectively determined according to a center point coordinate and a direction vector of the surface area matching the visible surface.
  2. 根据权利要求1所述的方法,其中,在获取包含目标实体的视频帧之前,还包括:The method of claim 1, wherein before acquiring the video frame including the target entity, the method further comprises:
    对目标实体进行预处理,构建所述目标实体的预制特征集合。Preprocessing the target entity to construct a set of prefabricated features of the target entity.
  3. 根据权利要求2所述的方法,其中,对目标实体进行预处理,构建所述目标实体的预制特征集合,包括:The method of claim 2, wherein pre-processing the target entity to construct the set of pre-made features of the target entity comprises:
    获取目标实体多个表面区域的图像,其中,所述多个表面区域至少覆盖所述目标实体的所有表面区域;Acquiring an image of a plurality of surface regions of the target entity, wherein the plurality of surface regions cover at least all surface regions of the target entity;
    对所述多个表面区域的图像进行图像特征提取,获取所述每个表面区域的图像特征信息;Performing image feature extraction on the images of the plurality of surface regions to acquire image feature information of each surface region;
    根据所述每个表面区域的图像特征信息构建所述目标实体的预制特征集合。Constructing a set of prefabricated features of the target entity based on the image feature information of each of the surface regions.
  4. 根据权利要求1至3中任一项所述的方法,其中,根据与所述可视表面匹配的表面区域的中心点坐标确定所述目标实体可视表面的中心点坐标,包括:The method according to any one of claims 1 to 3, wherein determining a center point coordinate of the visible surface of the target entity according to a center point coordinate of a surface area matching the visible surface comprises:
    若与所述可视表面匹配的表面区域有多个,将多个表面区域的中心点坐标进行加权平均处理,确定所述目标实体的中心点坐标。If there are a plurality of surface regions matching the visible surface, the center point coordinates of the plurality of surface regions are subjected to weighted averaging processing, and the center point coordinates of the target entity are determined.
  5. 根据权利要求1至3中任一项所述的方法,其中,根据与所述可视表面匹配的表面区域的方向矢量确定所述目标实体可视表面方向矢量,包括: The method according to any one of claims 1 to 3, wherein determining the target entity visible surface direction vector according to a direction vector of a surface area matching the visible surface comprises:
    若与所述可视表面匹配的表面区域有多个,将多个表面区域的方向矢量加权相加并进行矢量归一化处理,获取所述目标实体可视表面的方向矢量。If there are multiple surface regions matching the visible surface, the direction vectors of the plurality of surface regions are weighted and added and subjected to vector normalization processing to obtain a direction vector of the visible surface of the target entity.
  6. 一种实现增强现实的方法,其中,该方法包括:A method of implementing augmented reality, wherein the method comprises:
    采用所述方位跟踪方法获取所述目标实体的中心点坐标和方向矢量;Obtaining a center point coordinate and a direction vector of the target entity by using the azimuth tracking method;
    根据所述方向矢量由虚拟模型中获取与所述可视表面对应的显示部分;Obtaining a display portion corresponding to the visible surface from the virtual model according to the direction vector;
    根据所述中心点坐标将所述显示部分渲染至所述视频帧中,以使所述目标实体的可视表面由所述显示部分覆盖。The display portion is rendered into the video frame based on the center point coordinates such that a visible surface of the target entity is covered by the display portion.
    输出完成渲染的视频帧。Output the finished rendered video frame.
  7. 根据权利要求6所述的方法,其中,该方法还包括:The method of claim 6 wherein the method further comprises:
    获取所述显示部分中的信息叠加点;Obtaining an information superposition point in the display portion;
    将所述信息叠加点对应的虚拟信息渲染至所述视频帧中,以使所述显示部分的相应位置显示所述虚拟信息。And rendering virtual information corresponding to the information superposition point into the video frame, so that the corresponding position of the display part displays the virtual information.
  8. 根据权利要求6所述的方法,其中,所述虚拟模型设置有触发点;The method of claim 6, wherein the virtual model is provided with a trigger point;
    所述方法还包括:The method further includes:
    判断所述显示部分的触发点是否位于所述视频帧的触发区域;Determining whether a trigger point of the display portion is located in a trigger area of the video frame;
    若所述触发点处于所述视频帧的触发区域,将所述触发点的触发效果渲染至所述视频帧中。If the trigger point is in a trigger area of the video frame, the trigger effect of the trigger point is rendered into the video frame.
  9. 根据权利要求6至8中任一项所述的方法,其中,所述目标实体为实体地球仪,所述虚拟模型为与所述实体地球仪匹配的虚拟地球仪。A method according to any one of claims 6 to 8, wherein the target entity is a physical globe and the virtual model is a virtual globe that matches the physical globe.
  10. 一种目标实体的方位跟踪装置,其中,该装置包括:An orientation tracking device for a target entity, wherein the device comprises:
    图像采集模块,用于获取包含目标实体的视频帧;An image acquisition module, configured to acquire a video frame that includes a target entity;
    特征提取模块,用于对所述视频帧中的目标实体进行图像特征提取,获取所述目标实体可视表面的图像特征信息;a feature extraction module, configured to perform image feature extraction on a target entity in the video frame, and acquire image feature information of a visible surface of the target entity;
    特征匹配模块,用于基于所述目标实体的预制特征集合对获取到的所述目标实体可视表面的图像特征信息进行特征识别,在预制特征集合中获取与所述可视表面匹配的表面区域,其中,所述预制特征集合包含所述目标实体的多个表面区域以及每个表面区域的图像特征信息; a feature matching module, configured to perform feature recognition on the acquired image feature information of the visible surface of the target entity based on the set of prefabricated features of the target entity, and acquire a surface region matching the visible surface in the prefabricated feature set The set of prefabricated features includes a plurality of surface regions of the target entity and image feature information of each surface region;
    综合处理模块,用于根据与所述可视表面匹配的表面区域的中心点坐标和方向矢量分别确定所述目标实体可视表面的中心点坐标和方向矢量。And an integrated processing module, configured to respectively determine a center point coordinate and a direction vector of the visible surface of the target entity according to a center point coordinate and a direction vector of the surface area matching the visible surface.
  11. 根据权利要求10所述的装置,其中,该装置还包括:The device of claim 10, wherein the device further comprises:
    预处理模块,用于在获取包含目标实体的视频帧之前,对目标实体进行预处理,构建所述目标实体的预制特征集合。And a pre-processing module, configured to pre-process the target entity before acquiring the video frame that includes the target entity, and construct a pre-made feature set of the target entity.
  12. 根据权利要求11所述的装置,其中,所述预处理模块,用于控制所述图像采集模块获取目标实体多个表面区域的图像,其中,所述多个表面区域至少覆盖所述目标实体的所有表面区域;控制所述特征提取模块对所述多个表面区域的图像进行图像特征提取,获取所述每个表面区域的图像特征信息;以及根据所述每个表面区域的图像特征信息构建所述目标实体的预制特征集合。The apparatus according to claim 11, wherein the preprocessing module is configured to control the image acquisition module to acquire an image of a plurality of surface regions of a target entity, wherein the plurality of surface regions cover at least the target entity All the surface regions; controlling the feature extraction module to perform image feature extraction on the images of the plurality of surface regions, acquiring image feature information of each surface region; and constructing the image according to the image feature information of each surface region A set of prefabricated features of the target entity.
  13. 根据权利要求10至12中任一项所述的装置,其中,所述综合处理模块,用于在与所述可视表面匹配的表面区域有多个时,将多个表面区域的中心点坐标进行加权平均处理,确定所述目标实体的中心点坐标。The apparatus according to any one of claims 10 to 12, wherein the integrated processing module is configured to coordinate center points of the plurality of surface regions when there are a plurality of surface regions matching the visible surface A weighted averaging process is performed to determine the center point coordinates of the target entity.
  14. 根据权利要求10至12中任一项所述的装置,其中,所述综合处理模块,用于在与所述可视表面匹配的表面区域有多个时,将多个表面区域的方向矢量加权相加并进行矢量归一化处理,获取所述目标实体可视表面的方向矢量。The apparatus according to any one of claims 10 to 12, wherein the integrated processing module is configured to weight direction vectors of the plurality of surface regions when there are a plurality of surface regions matching the visible surface Adding and performing vector normalization processing to obtain a direction vector of the visible surface of the target entity.
  15. 一种实现增强现实的设备,其中,该设备包括:A device for realizing augmented reality, wherein the device comprises:
    方位跟踪装置,用于获取所述目标实体的中心点坐标和方向矢量;An orientation tracking device, configured to acquire a center point coordinate and a direction vector of the target entity;
    渲染装置,用于根据所述方向矢量由虚拟模型中获取与所述可视表面对应的显示部分,以及根据所述中心点坐标将所述显示部分渲染至所述视频帧中,以使所述目标实体的可视表面由所述显示部分覆盖;a rendering device, configured to acquire a display portion corresponding to the visible surface from the virtual model according to the direction vector, and render the display portion into the video frame according to the center point coordinate, so that the The visible surface of the target entity is covered by the display portion;
    输出装置,用于输出完成渲染的视频帧。An output device for outputting a finished video frame.
  16. 根据权利要求15所述的设备,其中,所述渲染装置,还用于获取所述显示部分中的信息叠加点,以及将所述信息叠加点对应的虚拟信息渲染至所述视频帧中,以使所述显示部分的相应位置显示所述虚拟信息。The device according to claim 15, wherein the rendering device is further configured to acquire an information superposition point in the display portion, and render virtual information corresponding to the information superimposition point into the video frame, The virtual information is displayed in a corresponding position of the display portion.
  17. 根据权利要求16所述的设备,其中,所述虚拟模型设置有触发点;The apparatus of claim 16, wherein the virtual model is provided with a trigger point;
    所述渲染装置,还用于判断所述显示部分的触发点是否位于所述视频 帧的触发区域,并在所述触发点处于所述视频帧的触发区域时,将所述触发点的触发效果渲染至所述视频帧中。The rendering device is further configured to determine whether a trigger point of the display portion is located in the video a triggering region of the frame, and when the triggering point is in a triggering region of the video frame, rendering a triggering effect of the triggering point into the video frame.
  18. 根据权利要求15至17中任一项所述的设备,其中,所述目标实体为实体地球仪,所述虚拟模型为与所述实体地球仪匹配的虚拟地球仪。 Apparatus according to any one of claims 15 to 17, wherein the target entity is a physical globe and the virtual model is a virtual globe that matches the physical globe.
PCT/CN2017/080276 2016-07-19 2017-04-12 Method and relevant apparatus for orientational tracking, method and device for realizing augmented reality WO2018014601A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201610570465.XA CN106251404B (en) 2016-07-19 2016-07-19 Orientation tracking, the method and relevant apparatus, equipment for realizing augmented reality
CN201610570465.X 2016-07-19

Publications (1)

Publication Number Publication Date
WO2018014601A1 true WO2018014601A1 (en) 2018-01-25

Family

ID=57613380

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/080276 WO2018014601A1 (en) 2016-07-19 2017-04-12 Method and relevant apparatus for orientational tracking, method and device for realizing augmented reality

Country Status (2)

Country Link
CN (1) CN106251404B (en)
WO (1) WO2018014601A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111093301A (en) * 2019-12-14 2020-05-01 安琦道尔(上海)环境规划建筑设计咨询有限公司 Light control method and system
CN113269810A (en) * 2018-04-11 2021-08-17 深圳市瑞立视多媒体科技有限公司 Motion gesture recognition method and device for catching ball
CN115057242A (en) * 2022-03-07 2022-09-16 上海箱云物流科技有限公司 Container management method and system based on combination of virtual reality and augmented reality

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106251404B (en) * 2016-07-19 2019-02-01 央数文化(上海)股份有限公司 Orientation tracking, the method and relevant apparatus, equipment for realizing augmented reality
CN108540824B (en) * 2018-05-15 2021-01-19 北京奇虎科技有限公司 Video rendering method and device
CN109582147B (en) * 2018-08-08 2022-04-26 亮风台(上海)信息科技有限公司 Method for presenting enhanced interactive content and user equipment
CN109656363B (en) * 2018-09-04 2022-04-15 亮风台(上海)信息科技有限公司 Method and equipment for setting enhanced interactive content
CN110084204B (en) * 2019-04-29 2020-11-24 北京字节跳动网络技术有限公司 Image processing method and device based on target object posture and electronic equipment
CN110335292B (en) * 2019-07-09 2021-04-30 北京猫眼视觉科技有限公司 Method, system and terminal for realizing simulation scene tracking based on picture tracking
CN110727350A (en) * 2019-10-09 2020-01-24 武汉幻石佳德数码科技有限公司 Augmented reality-based object identification method, terminal device and storage medium
CN112051746B (en) * 2020-08-05 2023-02-07 华为技术有限公司 Method and device for acquiring service
CN112565605B (en) * 2020-12-02 2022-11-25 维沃移动通信有限公司 Image display method and device and electronic equipment
CN117425917A (en) * 2021-03-08 2024-01-19 杭州他若定位科技有限公司 Scene triggering and interaction based on target positioning identification
CN118112606B (en) * 2024-04-30 2024-07-23 中孚安全技术有限公司 Accurate investigation system, method and medium for vehicle GPS tracker

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120243739A1 (en) * 2011-03-25 2012-09-27 Masaki Fukuchi Information processing device, object recognition method, program, and terminal device
CN103049729A (en) * 2012-12-30 2013-04-17 成都理想境界科技有限公司 Method, system and terminal for augmenting reality based on two-dimension code
CN104050475A (en) * 2014-06-19 2014-09-17 樊晓东 Reality augmenting system and method based on image feature matching
CN104966307A (en) * 2015-07-10 2015-10-07 成都品果科技有限公司 AR (augmented reality) algorithm based on real-time tracking
CN105069754A (en) * 2015-08-05 2015-11-18 意科赛特数码科技(江苏)有限公司 System and method for carrying out unmarked augmented reality on image
CN105338117A (en) * 2015-11-27 2016-02-17 亮风台(上海)信息科技有限公司 Method, device and system for generating AR applications and presenting AR instances
CN105491365A (en) * 2015-11-25 2016-04-13 罗军 Image processing method, device and system based on mobile terminal
CN106251404A (en) * 2016-07-19 2016-12-21 央数文化(上海)股份有限公司 Orientation tracking, the method realizing augmented reality and relevant apparatus, equipment

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102142055A (en) * 2011-04-07 2011-08-03 上海大学 True three-dimensional design method based on augmented reality interactive technology
US8855366B2 (en) * 2011-11-29 2014-10-07 Qualcomm Incorporated Tracking three-dimensional objects
CN103366610B (en) * 2013-07-03 2015-07-22 央数文化(上海)股份有限公司 Augmented-reality-based three-dimensional interactive learning system and method
CN104134229A (en) * 2014-08-08 2014-11-05 李成 Real-time interaction reality augmenting system and method
CN104463108B (en) * 2014-11-21 2018-07-31 山东大学 A kind of monocular real time target recognitio and pose measuring method

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120243739A1 (en) * 2011-03-25 2012-09-27 Masaki Fukuchi Information processing device, object recognition method, program, and terminal device
CN103049729A (en) * 2012-12-30 2013-04-17 成都理想境界科技有限公司 Method, system and terminal for augmenting reality based on two-dimension code
CN104050475A (en) * 2014-06-19 2014-09-17 樊晓东 Reality augmenting system and method based on image feature matching
CN104966307A (en) * 2015-07-10 2015-10-07 成都品果科技有限公司 AR (augmented reality) algorithm based on real-time tracking
CN105069754A (en) * 2015-08-05 2015-11-18 意科赛特数码科技(江苏)有限公司 System and method for carrying out unmarked augmented reality on image
CN105491365A (en) * 2015-11-25 2016-04-13 罗军 Image processing method, device and system based on mobile terminal
CN105338117A (en) * 2015-11-27 2016-02-17 亮风台(上海)信息科技有限公司 Method, device and system for generating AR applications and presenting AR instances
CN106251404A (en) * 2016-07-19 2016-12-21 央数文化(上海)股份有限公司 Orientation tracking, the method realizing augmented reality and relevant apparatus, equipment

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113269810A (en) * 2018-04-11 2021-08-17 深圳市瑞立视多媒体科技有限公司 Motion gesture recognition method and device for catching ball
CN111093301A (en) * 2019-12-14 2020-05-01 安琦道尔(上海)环境规划建筑设计咨询有限公司 Light control method and system
CN115057242A (en) * 2022-03-07 2022-09-16 上海箱云物流科技有限公司 Container management method and system based on combination of virtual reality and augmented reality
CN115057242B (en) * 2022-03-07 2023-08-01 上海箱云物流科技有限公司 Container management method based on combination of virtual reality and augmented reality

Also Published As

Publication number Publication date
CN106251404B (en) 2019-02-01
CN106251404A (en) 2016-12-21

Similar Documents

Publication Publication Date Title
WO2018014601A1 (en) Method and relevant apparatus for orientational tracking, method and device for realizing augmented reality
US11481923B2 (en) Relocalization method and apparatus in camera pose tracking process, device, and storage medium
CN109887003B (en) Method and equipment for carrying out three-dimensional tracking initialization
CN110910486B (en) Indoor scene illumination estimation model, method and device, storage medium and rendering method
US11748906B2 (en) Gaze point calculation method, apparatus and device
Ventura et al. Wide-area scene mapping for mobile visual tracking
JP2021526680A (en) Self-supervised training for depth estimation system
CN103839277B (en) A kind of mobile augmented reality register method of outdoor largescale natural scene
WO2019238114A1 (en) Three-dimensional dynamic model reconstruction method, apparatus and device, and storage medium
WO2023280038A1 (en) Method for constructing three-dimensional real-scene model, and related apparatus
CN115690382B (en) Training method of deep learning model, and method and device for generating panorama
WO2021093679A1 (en) Visual positioning method and device
CN112927362A (en) Map reconstruction method and device, computer readable medium and electronic device
CN112288878B (en) Augmented reality preview method and preview device, electronic equipment and storage medium
WO2022088819A1 (en) Video processing method, video processing apparatus and storage medium
CN112766215A (en) Face fusion method and device, electronic equipment and storage medium
CN112270709A (en) Map construction method and device, computer readable storage medium and electronic device
CN113724391A (en) Three-dimensional model construction method and device, electronic equipment and computer readable medium
US11100617B2 (en) Deep learning method and apparatus for automatic upright rectification of virtual reality content
CN114882106A (en) Pose determination method and device, equipment and medium
WO2022126921A1 (en) Panoramic picture detection method and device, terminal, and storage medium
CN110956571A (en) SLAM-based virtual-real fusion method and electronic equipment
CN108205820B (en) Plane reconstruction method, fusion method, device, equipment and storage medium
Zhou et al. Improved YOLOv7 models based on modulated deformable convolution and swin transformer for object detection in fisheye images
KR102176805B1 (en) System and method for providing virtual reality contents indicated view direction

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17830239

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17830239

Country of ref document: EP

Kind code of ref document: A1