WO2022121606A1 - Method and system for obtaining identification information of device or user thereof in scenario - Google Patents

Method and system for obtaining identification information of device or user thereof in scenario Download PDF

Info

Publication number
WO2022121606A1
WO2022121606A1 PCT/CN2021/129727 CN2021129727W WO2022121606A1 WO 2022121606 A1 WO2022121606 A1 WO 2022121606A1 CN 2021129727 W CN2021129727 W CN 2021129727W WO 2022121606 A1 WO2022121606 A1 WO 2022121606A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
user
camera
location information
spatial location
Prior art date
Application number
PCT/CN2021/129727
Other languages
French (fr)
Chinese (zh)
Inventor
方俊
李江亮
牛旭恒
Original Assignee
北京外号信息技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from CN202011440905.2A external-priority patent/CN112528699B/en
Priority claimed from CN202011440875.5A external-priority patent/CN112581630A/en
Priority claimed from CN202011442020.6A external-priority patent/CN114663491A/en
Application filed by 北京外号信息技术有限公司 filed Critical 北京外号信息技术有限公司
Publication of WO2022121606A1 publication Critical patent/WO2022121606A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/14Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods

Definitions

  • the present invention relates to the field of information interaction, and in particular, to a method and system for obtaining identification information of a device or its user in a scene.
  • sensors such as cameras and radars will be deployed in the scenario to sense, locate, and track the personnel or equipment that appears in the scenario.
  • these sensors can sense the position or movement of people or equipment in the scene, they cannot obtain identification information of these people or equipment, thus making it difficult to provide services for these people or equipment.
  • facial recognition technology can be used to identify people, this involves violating user privacy and may have legal risks.
  • these sensors can usually only realize one-way information transmission (that is, collect relevant information in the scene), and cannot provide information to the user in the scene based on this information (for example, the user's real-time location information), such as navigation information , instruction information, commercial promotion information, etc.
  • on-site manual service is usually adopted, which requires setting up some consultation desks and arranging service personnel at a certain density in the venue, which is costly and flexible. Low.
  • One aspect of the present invention relates to a method for obtaining identification information for a device or its user in a scene in which one or more sensors and one or more visual markers are deployed, the sensors being capable of being used for sensing measuring or determining the location information of the device or user in the scene, the method includes: receiving information sent by the device, the information including the identification information of the device or its user and the spatial location information of the device, wherein , the device determines its spatial location information by scanning the visual sign; identifies the device or its user within the sensing range of the sensor based on the spatial location information of the device; The identification information of its user is associated with the device or its user within the sensing range of the sensor in order to provide services to the device or its user.
  • Another aspect of the present invention relates to a system for obtaining identification information of a device in a scene or a user thereof, the system comprising: one or more sensors deployed in the scene, the sensors capable of sensing measuring or determining location information of devices or users in the scene; one or more visual markers deployed in the scene; and a server configured to implement the methods described in the embodiments of the present application.
  • Another aspect of the present invention relates to a storage medium, in which a computer program is stored, and when the computer program is executed by a processor, can be used to implement the method described in the embodiments of the present application.
  • Another aspect of the present invention relates to an electronic device, comprising a processor and a memory, wherein the memory stores a computer program, and when the computer program is executed by the processor, can be used to implement the described embodiments of the present application. method.
  • the solution of the present invention not only the position or movement of persons or equipment existing in the scene can be sensed, but also the identification information of these persons or equipment can be obtained, and the corresponding personnel or equipment can be provided with services through the identification information.
  • the location information of the user in the scene can be collected or monitored, but also information, such as navigation information, instruction information, business promotion information, etc., can be provided to the user based on the real-time location information of the user.
  • Figure 1 shows an exemplary visual sign
  • Figure 2 shows an optical communication device that can be used as a visual sign
  • Figure 3 illustrates a system for obtaining identification information of a device in a scene or its user, according to one embodiment
  • FIG. 4 illustrates a method for obtaining identification information of a device in a scene or a user thereof, according to one embodiment
  • Figure 5 illustrates a method for providing a service to a device in a scene or its user, according to one embodiment.
  • Figure 6 illustrates a method for providing information to a user in a scene through a device (here, glasses are used as an example), according to one embodiment
  • FIG. 7 illustrates a system for providing information to a user in a scene through glasses, according to one embodiment
  • Figure 8 illustrates a method for providing information to a user in a scene through glasses, according to one embodiment
  • Figure 9 illustrates a user interaction system according to one embodiment
  • Figure 10 illustrates a user interaction method according to one embodiment
  • FIG. 11 illustrates a first user and a virtual object associated with the first user as observed by a second user through his device, according to one embodiment
  • Figure 12 shows the actual image observed by a user through his cell phone screen, according to one embodiment.
  • Visual signs refer to signs that can be recognized by the human eye or electronic devices, which can have various forms.
  • visual markers may be used to convey information that can be obtained by smart devices (eg, cell phones, smart glasses, etc.).
  • the visual sign may be an optical communication device capable of emitting encoded optical information, or the visual sign may be a graphic with encoded information, such as a two-dimensional code (eg, QR code, applet code), barcode, or the like.
  • Figure 1 shows an exemplary visual sign with a specific black and white pattern.
  • FIG. 2 shows an optical communication device 100 that can be used as a visual sign, which includes three light sources (respectively, a first light source 101, a second light source 102, and a third light source 103).
  • the optical communication device 100 also includes a controller (not shown in FIG. 2 ) for selecting a corresponding driving mode for each light source according to the information to be communicated.
  • the controller can use different driving signals to control the light-emitting manner of the light source, so that when the optical communication apparatus 100 is photographed by using a device with an imaging function, the imaging of the light source can present different images. Appearance (eg, different colors, patterns, brightness, etc.).
  • the driving mode of each light source at the moment can be analyzed, so as to analyze the information transmitted by the optical communication device 100 at the moment.
  • each visual logo may be assigned an identification information (ID), which is used to uniquely identify or identify the visual logo by the manufacturer, manager or user of the visual logo, etc. .
  • ID an identification information
  • the user can use the device to capture the image of the visual sign to obtain the identification information transmitted by the visual sign, so as to access the corresponding service based on the identification information, for example, visit the webpage associated with the identification information, obtain the identification information associated with the identification information.
  • Other information eg, position or gesture information of the visual landmark corresponding to the identification information
  • the devices mentioned herein can be, for example, devices carried or controlled by users (eg, mobile phones, tablet computers, smart glasses, AR glasses, smart helmets, smart watches, cars, etc.), or machines that can move autonomously (eg, drones, driverless cars, robots, etc.).
  • the device can acquire the image containing the visual sign by collecting the image of the visual sign through the image acquisition device on it, and can identify the information transmitted by the visual sign and determine the relative position of the device relative to the visual sign by analyzing the imaging of the visual sign in the image. position or attitude information.
  • Sensors capable of sensing the location of objects may be various sensors capable of sensing or determining location information of objects in a scene, such as cameras, radars (eg, lidars, millimeter-wave radars), wireless signal transceivers, and the like.
  • a target in a scene can be a person or an object in the scene.
  • a camera is used as an example of a sensor for description.
  • Figure 3 illustrates a system for obtaining identification information of a device in a scene or its user, which can be used to provide services or information to a user in the scene through the device, according to one embodiment.
  • the system includes a visual sign 301, a camera 302, and a server (not shown in Figure 3).
  • User 303 is in the scene and carries device 304 .
  • the device 304 has an image capture device on it and can identify the visual sign 301 through the image capture device.
  • device 304 may be a cell phone carried by the user.
  • device 304 may be glasses worn by a user. The glasses themselves may have the ability to directly access the network, for example, the glasses may access the network by means of wifi, telecommunication network or the like.
  • the glasses may also not have the ability to directly access the network, but may indirectly access the network through a connection (eg, a Bluetooth connection or a wired connection) between it and the user's other devices (eg, a mobile phone, a watch, etc.).
  • a connection eg, a Bluetooth connection or a wired connection
  • the visual sign 301 and the camera 302 are each installed in the real scene in a specific position and attitude (hereinafter collectively referred to as "pose").
  • the server may obtain the respective pose information of the camera and the visual marker, and may obtain relative pose information between the camera and the visual marker based on the respective pose information of the camera and the visual marker.
  • the server may also directly obtain the relative pose information between the camera and the visual marker. In this way, the server can obtain a transformation matrix between the camera coordinate system and the visual sign coordinate system, and the transformation matrix may include, for example, a rotation matrix R and a displacement vector t between the two coordinate systems.
  • the camera may be a camera installed in a fixed position and having a fixed orientation, but it is understood that the camera may also be a camera that can move (for example, the position or direction can be changed), as long as its current pose information can be determined.
  • the current pose information of the camera can be set by the server, and the movement of the camera can be controlled based on the pose information, or the movement of the camera can be controlled by the camera itself or other devices, and the current pose information of the camera can be sent to the server.
  • more than one camera may be included in the system, and more than one visual sign may also be included.
  • a scene coordinate system (which may also be referred to as a real world coordinate system) may be established for the real scene, and the distance between the camera coordinate system and the scene coordinate system may be determined based on the pose information of the camera in the real scene and the transformation matrix between the visual landmark coordinate system and the scene coordinate system is determined based on the pose information of the visual landmark in the real scene.
  • the distance between the camera coordinate system and the scene coordinate system may be determined based on the pose information of the camera in the real scene
  • the transformation matrix between the visual landmark coordinate system and the scene coordinate system is determined based on the pose information of the visual landmark in the real scene.
  • having a relative pose between the camera and the visual sign refers to objectively having a relative pose between the two, and does not require the system to pre-store the relative pose information between the two above or Use this relative pose information.
  • the system may be stored in the system, and the relative poses of the two may not be calculated or used.
  • Cameras can be used to track objects in a real scene, which can be stationary or moving, such as people, stationary objects, movable objects, etc. in the scene.
  • a camera can be used to track the position of a person or object in a real scene by various methods in the prior art.
  • the location information of objects in the scene can be determined in combination with scene information (eg, information on the plane on which a person or object in the scene is located).
  • scene information eg, information on the plane on which a person or object in the scene is located.
  • the position information of the target can be determined according to the position of the target in the field of view of the camera and the depth information of the target.
  • the position information of the target can be determined according to the position of the target in the field of view of each camera.
  • the system may have multiple visual signs or multiple cameras, and the fields of view of the multiple cameras may be continuous or discontinuous.
  • Fig. 4 shows a method for obtaining identification information of a device in a scene or a user thereof according to an embodiment.
  • the method can be implemented using the system shown in Fig. 3 and can include the following steps:
  • Step 401 Receive information sent by the device, where the information includes identification information of the device or its user and spatial location information of the device.
  • the information sent by the device may be various information, such as alarm information, help information, service request information, and so on.
  • the identification information of the device or its user can be any information that can be used to identify or identify the device or its user, such as device ID information, the device's phone number, account information for an application on the device, the user's name or nickname, the user's identity information, user account information, etc.
  • the user 303 may use the device 304 to determine the spatial location information of the device 304 by scanning the visual markers 301 deployed in the scene.
  • the user 303 may send information to the server through the device 304, the information may include the spatial position information of the device 304, and the spatial position information may be the spatial position information of the device 304 relative to the visual sign 301 or the spatial position information of the device 304 in the scene .
  • an image of the visual sign 301 may be collected using the device 304; identification information of the visual sign 301 and spatial position information of the device 304 relative to the visual sign 301 are determined by analyzing the collected image of the visual sign 301; The identification information of 301 determines the position and attitude information of the visual marker 301 in space; and based on the position and attitude information of the visual marker 301 in space and the spatial position information of the device 304 relative to the visual marker 301, determine the position of the device 304 in the scene. Spatial location information.
  • the device 304 can send the identification information of the visual marker 301 and the spatial position information of the device 304 relative to the visual marker 301 to the server, so that the server can determine the spatial position information of the device 304 in the scene.
  • the device 304 can also be used to scan the visual marker 301 to determine the gesture information of the device 304 relative to the visual marker 301 or the gesture information of the device 304 in the scene, and the gesture information can be sent to the server.
  • the spatial position information and attitude information of the device may be the spatial position information and attitude information of the device when scanning the visual sign, or the real-time position information and attitude information at any moment after scanning the visual sign.
  • a device can determine its initial spatial position information and attitude information when scanning a visual sign, and then use various sensors built into the device (eg, acceleration sensor, magnetic sensor, orientation sensor, gravity sensor, gyroscope, camera, etc.)
  • the real-time position and/or attitude of the device is determined by measuring or tracking its position change and/or attitude change by methods known in the art (eg, inertial navigation, visual odometry, SLAM, VSLAM, SFM, etc.).
  • the spatial location information of the device received by the server may be coordinate information, but is not limited thereto, any information that can be used to derive the spatial location of the device belongs to spatial location information.
  • the spatial location information of the device received by the server may be an image of a visual sign captured by the device, and the server may determine the spatial location of the device according to the image.
  • any information that can be used to derive a device's pose is pose information, which in one embodiment may be an image of a visual landmark captured by the device.
  • Step 402 Identify the device or its user in the image captured by the camera based on the spatial location information of the device.
  • the imaging position of the device or its user in the image captured by the camera may be determined based on the spatial position information of the device, and the device or the user in the image captured by the camera may be identified according to the imaging position.
  • the imaging position of the user in the image captured by the camera can be determined based on the spatial location information of the device. Since the user usually scans the visual sign while holding the device or wearing the device, the spatial position of the user can be inferred according to the spatial position of the device, and then the imaging position of the user in the image captured by the camera can be determined according to the spatial position of the user. The imaging position of the device in the image captured by the camera can also be determined according to the spatial position of the device, and then the imaging position of the user can be inferred according to the imaging position of the device.
  • the imaging position of the device in the image captured by the camera can be determined based on the spatial location information of the device.
  • a pre-established mapping relationship between one or more spatial positions (not necessarily all) in the scene and one or more imaging positions in the image captured by the camera and the space of the device may be used Location information to determine the imaging position of the device or its user in the image captured by the camera. For example, for a hall scene, several spatial positions on the floor of the hall can be selected, and the imaging positions of these positions in the image captured by the camera can be determined. After that, the mapping relationship between these spatial positions and the imaging positions can be established, and the An imaging position corresponding to a certain spatial position is deduced based on the mapping relationship.
  • the imaging position of the device or its user in the image captured by the camera may be determined based on the spatial position information of the device and the pose information of the camera, where the pose information of the camera may be its position in the scene pose information or its pose information relative to visual landmarks.
  • the device or its user can be identified in the image according to the imaging position. For example, a device or user closest to the imaging position may be selected, or a device or user whose distance from the imaging position satisfies a predetermined condition may be selected.
  • the spatial location information of the device may be compared with the spatial location information of one or more devices or users determined according to the tracking result of the camera. Compare.
  • a camera can be used to determine the spatial position of a person or object in a real scene through various methods in the prior art. For example, in the case of using a single monocular camera, the location information of objects in the scene can be determined in combination with scene information (eg, information on the plane on which a person or object in the scene is located). For the case of using a binocular camera, the position information of the target can be determined according to the position of the target in the field of view of the camera and the depth information of the target.
  • the position information of the target can be determined according to the position of the target in the field of view of each camera.
  • the spatial location information of one or more users may also be determined by using images captured by a camera in combination with lidar and the like.
  • real-time spatial location information eg, satellite positioning information or location information obtained through the device's sensors
  • a camera tracks the location of the plurality of users or devices and identifies the device or a user thereof by comparing real-time spatial location information received from the device to the locations of the plurality of users or devices tracked by the camera.
  • feature information of the device user may be determined based on information sent by the device, and the multiple users may be collected by a camera.
  • feature information of a plurality of users and identify the device user by comparing the feature information of the plurality of users with the feature information of the device user.
  • the field of view can cover one or more cameras of the device or its user, and then the imaging position of the device or its user in the images captured by the one or more cameras is determined .
  • Step 403 Associate the identification information of the device or its user with the device or its user in the image captured by the camera, so as to use the identification information to provide a service to the device or its user.
  • the received identification information of the device or its user may be associated with the device or its user in the image.
  • the ID information, phone number, account information of an application on the device can be known, or the user's name or nickname, the user's identity information, the user's account information, and many more.
  • the identification information can be used to provide various services to the device or its user, such as navigation service, explanation service, information display service, and so on. In one embodiment, the above information may be provided visually, audibly, or the like.
  • a virtual object may be superimposed on a display medium of a device (eg, a mobile phone or glasses), and the virtual object may be, for example, an icon (eg, a navigation icon), a picture, a text, and the like.
  • a device eg, a mobile phone or glasses
  • the virtual object may be, for example, an icon (eg, a navigation icon), a picture, a text, and the like.
  • the steps in the method shown in FIG. 4 may be implemented by the server in the system shown in FIG. 3 , but it is understood that one or more of these steps may also be implemented by other devices.
  • the device or its user in the scene can also be tracked through a camera to obtain its real-time position information and/or attitude information, or the device can be used to obtain its real-time position information and/or attitude information .
  • services can be provided to the device or its user based on the location and/or attitude information.
  • information can be sent to the corresponding device or user in the field of view of the camera through the identification information , the information is, for example, navigation information, explanation information, instruction information, advertisement information, and so on.
  • One or more visual signs and one or more cameras are deployed in a smart factory scenario where robots are used to deliver goods.
  • the camera is used to track the position of the robot, and navigation instructions are sent to the robot according to the tracked position.
  • each robot may be made to scan a visual sign, for example, when entering the scene or the camera's field of view, and send its position information and identification information. In this way, the identification information of each robot within the field of view of the camera can be easily determined, so as to send each robot a travel instruction or a navigation instruction based on its current position and the work task to be completed.
  • information related to a virtual object may be sent to the device, the virtual object may be, for example, pictures, characters, numbers, icons, videos, three-dimensional models, etc., and the information related to the virtual object may include the spatial location of the virtual object information.
  • the virtual object can be presented on the display medium of the device.
  • the device may present the virtual object at an appropriate location on its display medium based on the device's or user's spatial location information and/or gesture information.
  • the virtual object may be presented on the display medium of the user equipment in an augmented reality or mixed reality manner, for example.
  • the virtual object is a video image or a dynamic three-dimensional model generated by video capture of live characters.
  • the virtual object may be a video image generated by real-time video capture of service personnel, and the video image may be presented on the display medium of the user equipment, so as to provide services to the user.
  • the spatial position of the video image can be set so that it can be presented on the display medium of the user equipment in the manner of augmented reality or mixed reality.
  • the identification information sent by the device or user within the field of view of the camera can be identified based on the identification information.
  • information such as service request information, alarm information, help information, comment information, and the like.
  • a virtual object associated with the device or the user may be set according to the information, wherein the spatial location information of the virtual object may be based on the information of the device or the user The position information of the virtual object can be determined, and the spatial position of the virtual object can be changed accordingly as the position of the device or the user changes.
  • the content of the virtual object may be updated according to new information received from the device or user (eg, a new comment by the user).
  • Fig. 5 shows a method for providing a service to a device or a user in a scene according to one embodiment.
  • the method can be implemented using the system shown in Fig. 3 and can include the following steps:
  • Step 501 Receive information sent by the device, where the information includes identification information of the device or its user and spatial location information of the device.
  • Step 502 Identify the device or its user in the image captured by the camera based on the spatial location information of the device.
  • Step 503 Mark the device or its user in the image captured by the camera.
  • the device or user can be identified using a variety of methods, for example, an image of the device or user can be framed, a particular icon can be presented adjacent to the device or user's image, or the device or user's image can be highlighted.
  • the imaging area of the marked device or user can be enlarged, or the camera can be made to shoot for the marked device or user.
  • the device or user can be continuously tracked through a camera, and real-time spatial location information and/or gesture information of the device or user can be determined.
  • Step 504 Associate the identification information of the device or its user with the device or its user in the image captured by the camera, so as to use the identification information to provide services to the device or its user.
  • a person who can observe the image captured by the camera can know that the device or user currently needs service, and can know that the device or user currently needs service.
  • the current location of the device or user so that various required services, such as explanation service, navigation service, consulting service, help service, etc., can be conveniently provided to the device or user.
  • the help desk deployed in the scenario can be replaced, and any user in the scenario can be provided with the services they need in a convenient and low-cost manner.
  • the service may be provided to the user through a device carried or controlled by the user, such as a mobile phone, smart glasses, a vehicle, and the like.
  • the service may be provided visually, audibly, etc. through a telephony function, an application (APP), etc. on the device.
  • APP application
  • the steps in the method shown in FIG. 5 may be implemented by the server in the system shown in FIG. 3 , but it is understood that one or more of these steps may also be implemented by other devices.
  • Fig. 6 shows a method for providing information to a user in a scene through a device (here, glasses are taken as an example) according to an embodiment, the method can be implemented using the system shown in Fig. 3, and can include the following steps :
  • Step 601 Receive information sent by the glasses, where the information includes spatial position information of the glasses.
  • the user may use the glasses to determine the spatial position information of the glasses by scanning the visual landmarks deployed in the scene.
  • the user can send information to the server through the glasses.
  • the glasses can also be used to scan the visual markers to determine the gesture information of the glasses relative to the visual markers or the gesture information of the glasses in the scene, and the gesture information can be sent to the server.
  • the information sent by the glasses may also include information related to the glasses or their users, such as service request information, help information, alarm information, identification information (such as phone numbers, APP account information) etc.
  • the glasses themselves may be capable of direct access to the network.
  • the glasses may not have the ability to directly access the network, but indirectly access the network through a connection between it and, for example, the user's mobile phone.
  • the server may use an intermediate device such as a mobile phone. Receive information sent by glasses.
  • Step 602 Identify the user of the glasses in the image captured by the camera based on the spatial position information of the glasses.
  • the user's identification information can be associated with the user in order to provide services to the user using the identification information.
  • Step 603 Track the user through the camera and update the spatial location information of the user.
  • a camera may be used to track the user and update the imaging position of the user, and determine the spatial position information of the user based on the updated imaging position.
  • Various visual tracking methods known in the art can be used to track the user in the field of view of the camera and update the imaging position of the user.
  • the camera can remain stationary or move while tracking the user.
  • multiple cameras may be used, which may have a continuous field of view or a discontinuous field of view. Where the field of view is discontinuous, the user's characteristics can be recorded and re-identified and tracked when the user re-enters the field of view of one or more cameras.
  • a pre-established mapping relationship between one or more spatial positions (not necessarily all) in the scene and one or more imaging positions in the image captured by the camera and the imaging positions may be used, to determine the user's spatial location information.
  • the spatial position information of the user may be determined based on the pose information of the camera and the imaging position. For example, in the case of using a depth camera or a multi-camera camera, the direction of the user relative to the camera can be determined based on the imaging position, the depth information can be used to determine the distance of the user relative to the camera, so as to determine the position of the user relative to the camera, and then , the spatial position information of the user can be further determined based on the pose information of the camera.
  • the distance of the user relative to the camera may be estimated based on the imaging size of the user, and the spatial position information of the user may be determined based on the pose information of the camera and the imaging position.
  • the distance of the user relative to the camera may be determined by using a lidar or the like installed on the camera, and the spatial position information of the user may be determined based on the pose information of the camera and the imaging position.
  • the multiple cameras can be used to jointly determine the spatial location information of the user.
  • the spatial position information of the user may be determined based on the pose information of the camera, the imaging position, and optional other information (eg, coordinate information of the ground in the scene).
  • the user's gesture information may also be determined based on the tracking result of the user by the camera.
  • Step 604 Provide information to the user through the user's glasses based on the user's spatial location information.
  • the user can be provided with various required information, such as navigation information, instruction information, tutorial information, advertising information, other information related to location-based services, and the like.
  • the above information may be provided visually, audibly, or the like.
  • a virtual object may be superimposed on the display medium of the glasses, and the virtual object may be, for example, an icon (eg, a navigation icon), a picture, a text, or the like.
  • the glasses themselves may have the ability to directly access the network, so that the glasses may directly receive indication information from the server.
  • the glasses may not have the ability to directly access the network, but indirectly access the network through a connection between it and, for example, the user's mobile phone, in this case, the glasses may pass through an intermediate device such as a mobile phone Receive instructions from the server.
  • information may be further provided to the user in conjunction with the glasses or the gesture information of the user thereof.
  • the posture information of the glasses or the user thereof may be determined by the glasses, or the posture information of the user may be determined by the user image captured by the camera, and the posture information may include the orientation information of the user.
  • the posture information of the glasses can be obtained through its built-in sensors, for example, by tracking the initial posture or directly determined by the built-in sensors of the glasses (for example, a gravity sensor, a magnetic sensor, an orientation sensor, etc.) .
  • the server may directly receive the gesture information from the glasses, or receive the gesture information through an intermediate device such as a mobile phone.
  • the steps in the method shown in FIG. 6 may be implemented by the server in the system shown in FIG. 3 , but it is understood that one or more of these steps may also be implemented by other devices.
  • FIG. 7 shows a system for providing information to a user in a scene through glasses, including a visual sign 701, a camera 702, and a server (not shown in FIG. 7), according to one embodiment.
  • a user 703 is in the scene and carries glasses 704 and a mobile phone 705 .
  • the mobile phone 705 can recognize the visual sign 701 through the image capture device on it, so the glasses 704 may not have an image capture device, or although there is an image capture device on it, the image capture device may not have the ability to recognize the visual sign 701 .
  • FIG. 8 illustrates a method of providing information to a user in a scene through glasses, which may be implemented using the system shown in FIG. 7 , according to one embodiment.
  • the method includes the following steps (part of the steps are similar to the steps in FIG. 6 , and will not be repeated here, but it can be understood that the content described for each step in FIG. 6 can also be applied to the corresponding steps in FIG. 8 ):
  • Step 801 Receive information sent by the user's mobile phone, where the information includes spatial location information of the mobile phone.
  • the user can use the mobile phone to determine the spatial location information of the mobile phone by scanning the visual landmarks deployed in the scene.
  • the gesture information of the mobile phone can also be determined by scanning the visual sign, and the gesture information can be sent to the server.
  • Step 802 Identify the user of the mobile phone in the image captured by the camera based on the spatial location information of the mobile phone.
  • the user's identification information can be associated with the user in order to provide services to the user using the identification information.
  • Step 803 Track the user through the camera and update the spatial location information of the user.
  • the user's gesture information can also be determined.
  • Step 804 Provide information to the user through the user's glasses based on the user's spatial location information.
  • the glasses themselves may have the ability to directly access the network, so that the glasses may directly receive indication information from the server.
  • the glasses may not have the ability to directly access the network, but indirectly access the network through a connection between it and, for example, the user's mobile phone, in this case, the glasses may pass through an intermediate device such as a mobile phone Receive instructions from the server.
  • the server may first send first information to the user's mobile phone, and then the mobile phone may send second information (the second information may be the same as or different from the first information) to the glasses based on the first information, so as to provide the user with information based on the glasses through the glasses. location services.
  • information may be further provided to the user in conjunction with the glasses or the gesture information of the user thereof.
  • the user may also not use the glasses, but only use the cell phone.
  • information may be provided to the user through the user's mobile phone based on the user's spatial location information.
  • the information may be further provided to the user in combination with the gesture information of the mobile phone or its user.
  • the gesture information of the user can be determined through the mobile phone, or the user's gesture information can be determined through the user image captured by the camera.
  • the gesture information of the mobile phone can be obtained through its built-in sensor.
  • a device for scanning a visual sign to determine its spatial location information may be referred to as a "position acquisition device", and a device for providing information to a user may be referred to as an "information receiving device”.
  • the location obtaining device and the information receiving device may be the same device, such as the user's mobile phone or the user's glasses; the location obtaining device and the information receiving device may also be different devices, such as a mobile phone and a user's glasses respectively. Glasses.
  • Figure 9 illustrates a user interaction system including a visual sign 901, a camera 902, and a server (not shown in Figure 9) according to one embodiment.
  • the camera and the visual sign are each deployed in a real scene with a specific position and attitude (hereinafter collectively referred to as "pose"), and the scene also has a first user 903 and a second user 905 who carry a first device 904 respectively. and the second device 906.
  • the first device 904 and the second device 906 have image capture devices on them and can identify the visual sign 901 through the image capture devices.
  • the first device 904 and the second device 906 may be, for example, mobile phones, glasses and other devices.
  • FIG. 10 shows a user interaction method according to one embodiment, which can be implemented using the above-mentioned system, and can include the following steps:
  • Step 1001 Receive information sent by a first device of a first user, where the information includes spatial location information of the first device and identification information of the first user or the first device.
  • the first user may use the first device to determine the spatial location information of the first device by scanning the visual markers deployed in the scene.
  • the first device can also be used to scan the visual marker to determine the gesture information of the first device relative to the visual marker or the gesture information of the first device in the scene, and the gesture information can be sent to the server.
  • Step 1002 Identify the first user in the image captured by the camera based on the spatial location information of the first device.
  • Step 1003 Associate the identification information of the first user or the first device with the first user in the image captured by the camera.
  • Step 1004 Track the first user through the camera and update the spatial location information of the first user.
  • the gesture information of the user or the device may also be determined based on the tracking result of the user or the device by the camera.
  • Step 1005 Set relevant information of the first virtual object associated with the first user, the relevant information includes content information and spatial position information, wherein the first virtual object is set according to the spatial position information of the first user. Spatial location information.
  • the spatial position of the first virtual object may be configured to be at a predetermined distance above the first user.
  • the content information of the first virtual object is related information used to describe the content of the virtual object, which may include, for example, pictures, characters, numbers, icons, animations, videos, three-dimensional models, etc. contained in the virtual object, and may also include virtual objects shape information, color information, size information, posture information, etc.
  • the content information of the first virtual object may be set according to the information from the first user or the first device identified by the identification information of the first user or the first device.
  • the content information of the first virtual object may be, for example, the occupation, identity, gender, age, name, nickname, etc. of the first user.
  • the spatial location information of the first virtual object may change accordingly as the location of the first user changes, and the virtual object may be updated according to new information received from the first user or the first device (eg, new comments by the user).
  • Content information of the object eg, updating the textual content of the virtual object.
  • the pose information of the virtual object may also be set, and the pose information of the virtual object may be set based on the pose information of the device or user associated therewith, but may also be set in other ways.
  • Step 1006 Send the relevant information of the first virtual object to the second device of the second user.
  • Information about the first virtual object can be used by the second device to render the first virtual object on its display medium based on its position information and/or gesture information (eg, in an augmented or mixed reality manner) .
  • the location information and attitude information of the second device may be determined in various feasible ways.
  • the second device may determine its position information and/or gesture information by scanning the visual landmarks.
  • the location information and/or posture information of the second device may be determined through the tracking result of the second device or its user by the camera.
  • the second device may also use various sensors built in it to determine its position information and/or attitude information.
  • the second device may use point cloud information of the scene to determine its position information and/or pose information.
  • the first virtual object after obtaining the spatial position information of the first virtual object and the position and attitude information of the second device, the first virtual object can be superimposed at a suitable position in the real scene presented by the display medium of the second device virtual object.
  • the gesture of the superimposed first virtual object may be further determined.
  • the user of the second device may perform various interactive operations on the first virtual object.
  • a second virtual object may also be set for the second user of the second device in a similar manner, and the content information and spatial location information of the second virtual object may be sent to the first device or the first device of the first user.
  • the first device and the other devices may be, for example, a mobile phone and glasses, respectively
  • the content information and the spatial location information of the second virtual object can be used by the first device or the other devices to be based on its location information and/or gesture information to present the second virtual object on its display medium.
  • the steps in the method shown in FIG. 10 may be implemented by the server in the system shown in FIG. 9 , but it is understood that one or more of these steps may also be implemented by other devices.
  • the virtual object may be, for example, an icon containing text, wherein the text is "pick-up, XXX of XX company".
  • the spatial position of the virtual object is associated with the spatial position of the first user and can move as the first user moves.
  • Figure 12 shows an actual image observed by a user through his cell phone screen, the image including multiple users, each user having an associated virtual object, according to one embodiment.
  • a camera is used as an example of a sensor for description, but it can be understood that the embodiments herein are also applicable to any other sensor that can sense or determine the target position, such as lidar, millimeter-wave radar, wireless Signal Transceivers, etc.
  • the devices involved in the embodiments of the present application may be any devices carried or controlled by the user (eg, mobile phones, tablet computers, smart glasses, AR glasses, smart helmets, smart watches, vehicles, etc.), and also It can be various machines that can move autonomously, for example, unmanned aerial vehicles, unmanned vehicles, robots, etc., and image acquisition devices are installed on the equipment.
  • the glasses in this application may be AR glasses, smart glasses, or any other glasses that can be used to present information to the user.
  • the glasses in this application also include glasses formed by adding components or inserts to ordinary optical glasses, for example, glasses formed by adding a display device to ordinary optical glasses.
  • the present invention may be implemented in the form of a computer program.
  • the computer program can be stored in various storage media (eg, hard disk, optical disk, flash memory, etc.), and when the computer program is executed by the processor, can be used to implement the method of the present invention.
  • the present invention may be implemented in the form of an electronic device.
  • the electronic device includes a processor and a memory, and the memory stores a computer program that, when executed by the processor, can be used to implement the method of the present invention.
  • references herein to "various embodiments,” “some embodiments,” “one embodiment,” or “an embodiment” etc. refer to the fact that a particular feature, structure, or property described in connection with the embodiment is included in the in at least one embodiment.
  • appearances of the phrases “in various embodiments,” “in some embodiments,” “in one embodiment,” or “in an embodiment” in various places throughout this document are not necessarily referring to the same implementation example.
  • the particular features, structures, or properties may be combined in any suitable manner in one or more embodiments.
  • particular features, structures, or properties shown or described in connection with one embodiment may be combined, in whole or in part, with the features, structures, or properties of one or more other embodiments without limitation, so long as the combination does not limit the Logical or not working.

Abstract

Provided are a method and system for obtaining identification information of a device or a user thereof in a scenario. One or more sensors and one or more visual marks are deployed in the scenario, and the sensor can be used for sensing or determining position information of a device or a user in the scenario. The method comprises: receiving information sent by a device, wherein the information comprises identification information of the device or a user thereof and spatial position information of the device, and the device determines the spatial position information thereof by scanning a visual mark; identifying the device or the user thereof within a sensing range of a sensor on the basis of the spatial position information of the device; and associating the identification information of the device or the user thereof with the device or the user thereof within the sensing range of the sensor, so as to provide a service for the device or the user thereof.

Description

用于获得场景中的设备或其用户的标识信息的方法和系统Method and system for obtaining identification information of a device in a scene or its user 技术领域technical field
本发明涉及信息交互领域,尤其涉及一种用于获得场景中的设备或其用户的标识信息的方法和系统。The present invention relates to the field of information interaction, and in particular, to a method and system for obtaining identification information of a device or its user in a scene.
背景技术Background technique
本部分的陈述仅仅是为了提供与本申请的技术方案有关的背景信息,以帮助理解,其对于本申请的技术方案而言并不一定构成现有技术。The statements in this section are only for providing background information related to the technical solutions of the present application to help understanding, and they do not necessarily constitute prior art to the technical solutions of the present application.
在许多场景中,基于安全、监控、公共服务等需要,会在场景中部署摄像头、雷达等传感器,以对场景中出现的人员或设备进行感测、定位、跟踪等。然而,这些传感器虽然能够感测到场景中存在的人员或设备的位置或移动,但并不能获得这些人员或设备的标识信息,因此,这使得难以为这些人员或设备提供服务。尽管可以使用人脸识别技术来识别人员信息,但这会涉及侵犯用户隐私,可能存在法律风险。另外,这些传感器通常仅能实现单向的信息传输(也即,采集场景中的相关信息),并不能基于这些信息(例如,用户的实时位置信息)向场景中的用户提供信息,例如导航信息、指示信息、商业推广信息等。在现有技术中,为了向场景中的用户提供服务,通常采用现场人工服务的方式,这需要在场所中以一定的密度设置一些咨询台并安排服务人员,这种方式成本较高并且灵活度低。In many scenarios, based on the needs of security, surveillance, and public services, sensors such as cameras and radars will be deployed in the scenario to sense, locate, and track the personnel or equipment that appears in the scenario. However, although these sensors can sense the position or movement of people or equipment in the scene, they cannot obtain identification information of these people or equipment, thus making it difficult to provide services for these people or equipment. Although facial recognition technology can be used to identify people, this involves violating user privacy and may have legal risks. In addition, these sensors can usually only realize one-way information transmission (that is, collect relevant information in the scene), and cannot provide information to the user in the scene based on this information (for example, the user's real-time location information), such as navigation information , instruction information, commercial promotion information, etc. In the prior art, in order to provide services to users in the scene, on-site manual service is usually adopted, which requires setting up some consultation desks and arranging service personnel at a certain density in the venue, which is costly and flexible. Low.
发明内容SUMMARY OF THE INVENTION
本发明的一个方面涉及一种用于获得场景中的设备或其用户的标识信息的方法,所述场景中部署有一个或多个传感器和一个或多个视觉标志,所述传感器能够用于感测或确定所述场景中的设备或用户的位置信息,所述方法包括:接收由设备发送的信息,该信息中包括所述设备或其用户的标识信息以及所述设备的空间位置信息,其中,所述设备通过扫描所述视觉标志来确定其空间位置信息;基于所述设备的空间位置信息识别 出在所述传感器的感测范围内的所述设备或其用户;以及将所述设备或其用户的标识信息关联到在所述传感器的感测范围内的所述设备或其用户,以便向所述设备或其用户提供服务。One aspect of the present invention relates to a method for obtaining identification information for a device or its user in a scene in which one or more sensors and one or more visual markers are deployed, the sensors being capable of being used for sensing measuring or determining the location information of the device or user in the scene, the method includes: receiving information sent by the device, the information including the identification information of the device or its user and the spatial location information of the device, wherein , the device determines its spatial location information by scanning the visual sign; identifies the device or its user within the sensing range of the sensor based on the spatial location information of the device; The identification information of its user is associated with the device or its user within the sensing range of the sensor in order to provide services to the device or its user.
本发明的另一个方面涉及一种用于获得场景中的设备或其用户的标识信息的系统,所述系统包括:部署于所述场景中的一个或多个传感器,所述传感器能够用于感测或确定所述场景中的设备或用户的位置信息;部署于所述场景中的一个或多个视觉标志;以及服务器,其被配置用于实现本申请的实施例所描述的方法。Another aspect of the present invention relates to a system for obtaining identification information of a device in a scene or a user thereof, the system comprising: one or more sensors deployed in the scene, the sensors capable of sensing measuring or determining location information of devices or users in the scene; one or more visual markers deployed in the scene; and a server configured to implement the methods described in the embodiments of the present application.
本发明的另一个方面涉及一种存储介质,其中存储有计算机程序,在所述计算机程序被处理器执行时,能够用于实现本申请的实施例所描述的方法。Another aspect of the present invention relates to a storage medium, in which a computer program is stored, and when the computer program is executed by a processor, can be used to implement the method described in the embodiments of the present application.
本发明的另一个方面涉及一种电子设备,包括处理器和存储器,所述存储器中存储有计算机程序,在所述计算机程序被处理器执行时,能够用于实现本申请的实施例所描述的方法。Another aspect of the present invention relates to an electronic device, comprising a processor and a memory, wherein the memory stores a computer program, and when the computer program is executed by the processor, can be used to implement the described embodiments of the present application. method.
通过本发明的方案,不仅能够感测场景中存在的人员或设备的位置或移动,还可以获得这些人员或设备的标识信息,并通过该标识信息为相应的人员或设备提供服务。另外,在一些实施例中,不仅能够采集或监视场景中用户的位置信息,而且可以基于用户的实时位置信息向用户提供信息,例如导航信息、指示信息、商业推广信息等。Through the solution of the present invention, not only the position or movement of persons or equipment existing in the scene can be sensed, but also the identification information of these persons or equipment can be obtained, and the corresponding personnel or equipment can be provided with services through the identification information. In addition, in some embodiments, not only the location information of the user in the scene can be collected or monitored, but also information, such as navigation information, instruction information, business promotion information, etc., can be provided to the user based on the real-time location information of the user.
附图说明Description of drawings
以下参照附图对本发明的实施例作进一步说明,其中:Embodiments of the present invention are further described below with reference to the accompanying drawings, wherein:
图1示出了一种示例性的视觉标志;Figure 1 shows an exemplary visual sign;
图2示出了一种可以作为视觉标志的光通信装置;Figure 2 shows an optical communication device that can be used as a visual sign;
图3示出了根据一个实施例的用于获得场景中的设备或其用户的标识信息的系统;Figure 3 illustrates a system for obtaining identification information of a device in a scene or its user, according to one embodiment;
图4示出了根据一个实施例的用于获得场景中的设备或其用户的标识信息的方法;4 illustrates a method for obtaining identification information of a device in a scene or a user thereof, according to one embodiment;
图5示出了根据一个实施例的用于向场景中的设备或其用户提供服务的方法。Figure 5 illustrates a method for providing a service to a device in a scene or its user, according to one embodiment.
图6示出了根据一个实施例的用于通过设备(在此以眼镜为例)向场 景中的用户提供信息的方法;Figure 6 illustrates a method for providing information to a user in a scene through a device (here, glasses are used as an example), according to one embodiment;
图7示出了根据一个实施例的用于通过眼镜向场景中的用户提供信息的系统;7 illustrates a system for providing information to a user in a scene through glasses, according to one embodiment;
图8示出了根据一个实施例的用于通过眼镜向场景中的用户提供信息的方法;Figure 8 illustrates a method for providing information to a user in a scene through glasses, according to one embodiment;
图9示出了根据一个实施例的用户交互系统;Figure 9 illustrates a user interaction system according to one embodiment;
图10示出了根据一个实施例的用户交互方法;Figure 10 illustrates a user interaction method according to one embodiment;
图11示出了根据一个实施例的由第二用户通过其设备观察到的第一用户以及与该第一用户关联的虚拟对象;11 illustrates a first user and a virtual object associated with the first user as observed by a second user through his device, according to one embodiment;
图12示出了根据一个实施例的由一个用户通过其手机屏幕观察到的实际图像。Figure 12 shows the actual image observed by a user through his cell phone screen, according to one embodiment.
具体实施方式Detailed ways
为了使本发明的目的、技术方案及优点更加清楚明白,以下结合附图通过具体实施例对本发明进一步详细说明。应当理解,此处所描述的具体实施例仅仅用以解释本发明,并不用于限定本发明。In order to make the objectives, technical solutions and advantages of the present invention clearer, the present invention will be further described in detail below with reference to the accompanying drawings through specific embodiments. It should be understood that the specific embodiments described herein are only used to explain the present invention, but not to limit the present invention.
视觉标志是指能够被人眼或者电子设备识别的标志,其可以具有各种各样的形式。在一些实施例中,视觉标志可以用于传递信息,该信息能够被智能设备(例如手机、智能眼镜等)获得。例如,视觉标志可以是能够发出编码的光信息的光通信装置,或者视觉标志可以是带编码信息的图形,例如二维码(例如QR码、小程序码)、条形码等。图1示出了一种示例性的视觉标志,其具有特定的黑白图案。图2示出了一种可以作为视觉标志的光通信装置100,其包括三个光源(分别是第一光源101、第二光源102、第三光源103)。光通信装置100还包括控制器(在图2中未示出),其用于根据要传递的信息为每个光源选择相应的驱动模式。例如,在不同的驱动模式下,控制器可以使用不同的驱动信号来控制光源的发光方式,从而使得当使用具有成像功能的设备拍摄光通信装置100时,其中的光源的成像可以呈现出不同的外观(例如,不同的颜色、图案、亮度、等等)。通过分析光通信装置100中的光源的成像,可以解析出各个光源此刻的驱动模式,从而解析出光通信装置100此刻传递的信息。Visual signs refer to signs that can be recognized by the human eye or electronic devices, which can have various forms. In some embodiments, visual markers may be used to convey information that can be obtained by smart devices (eg, cell phones, smart glasses, etc.). For example, the visual sign may be an optical communication device capable of emitting encoded optical information, or the visual sign may be a graphic with encoded information, such as a two-dimensional code (eg, QR code, applet code), barcode, or the like. Figure 1 shows an exemplary visual sign with a specific black and white pattern. FIG. 2 shows an optical communication device 100 that can be used as a visual sign, which includes three light sources (respectively, a first light source 101, a second light source 102, and a third light source 103). The optical communication device 100 also includes a controller (not shown in FIG. 2 ) for selecting a corresponding driving mode for each light source according to the information to be communicated. For example, in different driving modes, the controller can use different driving signals to control the light-emitting manner of the light source, so that when the optical communication apparatus 100 is photographed by using a device with an imaging function, the imaging of the light source can present different images. Appearance (eg, different colors, patterns, brightness, etc.). By analyzing the imaging of the light sources in the optical communication device 100 , the driving mode of each light source at the moment can be analyzed, so as to analyze the information transmitted by the optical communication device 100 at the moment.
为了基于视觉标志向用户提供相应的服务,每个视觉标志可以被分配一个标识信息(ID),该标识信息用于由视觉标志的制造者、管理者或使用者等唯一地识别或标识视觉标志。用户可以使用设备对视觉标志进行图像采集来获得该视觉标志传递的标识信息,从而可以基于该标识信息来访问相应的服务,例如,访问与标识信息相关联的网页、获取与标识信息相关联的其他信息(例如,与该标识信息对应的视觉标志的位置或姿态信息)等等。本文提到的设备例如可以是用户携带或控制的设备(例如,手机、平板电脑、智能眼镜、AR眼镜、智能头盔、智能手表、汽车等等),也可以是能够自主移动的机器(例如,无人机、无人驾驶汽车、机器人等等)。设备可以通过其上的图像采集器件对视觉标志进行图像采集来获得包含视觉标志的图像,并可以通过分析图像中的视觉标志的成像来识别出视觉标志传递的信息以及确定设备相对于视觉标志的位置或姿态信息。In order to provide corresponding services to users based on the visual logo, each visual logo may be assigned an identification information (ID), which is used to uniquely identify or identify the visual logo by the manufacturer, manager or user of the visual logo, etc. . The user can use the device to capture the image of the visual sign to obtain the identification information transmitted by the visual sign, so as to access the corresponding service based on the identification information, for example, visit the webpage associated with the identification information, obtain the identification information associated with the identification information. Other information (eg, position or gesture information of the visual landmark corresponding to the identification information) and so on. The devices mentioned herein can be, for example, devices carried or controlled by users (eg, mobile phones, tablet computers, smart glasses, AR glasses, smart helmets, smart watches, cars, etc.), or machines that can move autonomously (eg, drones, driverless cars, robots, etc.). The device can acquire the image containing the visual sign by collecting the image of the visual sign through the image acquisition device on it, and can identify the information transmitted by the visual sign and determine the relative position of the device relative to the visual sign by analyzing the imaging of the visual sign in the image. position or attitude information.
能够感测目标位置的传感器可以是各种能够用于感测或确定场景中的目标的位置信息的传感器,例如摄像头、雷达(例如激光雷达、毫米波雷达)、无线信号收发器、等等。场景中的目标可以是场景中的人或物体。在下文的实施例中,以摄像头作为传感器的示例进行描述。Sensors capable of sensing the location of objects may be various sensors capable of sensing or determining location information of objects in a scene, such as cameras, radars (eg, lidars, millimeter-wave radars), wireless signal transceivers, and the like. A target in a scene can be a person or an object in the scene. In the following embodiments, a camera is used as an example of a sensor for description.
图3示出了根据一个实施例的用于获得场景中的设备或其用户的标识信息的系统,该系统可以用于通过设备向场景中的用户提供服务或信息。该系统包括视觉标志301、摄像头302以及服务器(图3中未示出)。用户303位于场景中,并且携带有设备304。设备304上具有图像采集器件并能够通过图像采集器件识别视觉标志301。在一个实施例中,设备304可以是用户携带的手机。在一个实施例中,设备304可以是用户佩戴的眼镜。眼镜本身可以具备直接接入网络的能力,其例如可以通过wifi、电信网络等方式来接入网络。眼镜也可以不具备直接接入网络的能力,而是可以通过其与用户的其他设备(例如手机、手表等)之间的连接(例如蓝牙连接或有线连接)来间接地接入网络。Figure 3 illustrates a system for obtaining identification information of a device in a scene or its user, which can be used to provide services or information to a user in the scene through the device, according to one embodiment. The system includes a visual sign 301, a camera 302, and a server (not shown in Figure 3). User 303 is in the scene and carries device 304 . The device 304 has an image capture device on it and can identify the visual sign 301 through the image capture device. In one embodiment, device 304 may be a cell phone carried by the user. In one embodiment, device 304 may be glasses worn by a user. The glasses themselves may have the ability to directly access the network, for example, the glasses may access the network by means of wifi, telecommunication network or the like. The glasses may also not have the ability to directly access the network, but may indirectly access the network through a connection (eg, a Bluetooth connection or a wired connection) between it and the user's other devices (eg, a mobile phone, a watch, etc.).
视觉标志301和摄像头302各自以特定的位置和姿态(下文中可以统称为“位姿”)安装于现实场景中。在一个实施例中,服务器可以获得摄像头和视觉标志各自的位姿信息,并且可以基于摄像头和视觉标志各自的位姿信息得到摄像头和视觉标志之间的相对位姿信息。在一个实施例中, 服务器也可以直接得到摄像头和视觉标志之间的相对位姿信息。如此,服务器可以获得摄像头坐标系和视觉标志坐标系之间的变换矩阵,该变换矩阵例如可以包括两个坐标系之间的旋转矩阵R和位移向量t。通过摄像头坐标系和视觉标志坐标系之间的变换矩阵,可以将一个坐标系中的坐标转换为另一个坐标系中的坐标。摄像头可以是安装于固定位置并且具有固定朝向的摄像头,但可以理解,摄像头也可以是可移动(例如,可以改变位置或调整方向)的摄像头,只要能够确定其当前位姿信息即可。可以由服务器设置摄像头的当前位姿信息,并基于该位姿信息控制摄像头的移动,也可以由摄像头本身或者其他装置来控制摄像头的移动,并将摄像头的当前位姿信息发送给服务器。在一些实施例中,系统中可以包括不止一个摄像头,也可以包括不止一个视觉标志。The visual sign 301 and the camera 302 are each installed in the real scene in a specific position and attitude (hereinafter collectively referred to as "pose"). In one embodiment, the server may obtain the respective pose information of the camera and the visual marker, and may obtain relative pose information between the camera and the visual marker based on the respective pose information of the camera and the visual marker. In one embodiment, the server may also directly obtain the relative pose information between the camera and the visual marker. In this way, the server can obtain a transformation matrix between the camera coordinate system and the visual sign coordinate system, and the transformation matrix may include, for example, a rotation matrix R and a displacement vector t between the two coordinate systems. Through the transformation matrix between the camera coordinate system and the visual marker coordinate system, the coordinates in one coordinate system can be converted into the coordinates in the other coordinate system. The camera may be a camera installed in a fixed position and having a fixed orientation, but it is understood that the camera may also be a camera that can move (for example, the position or direction can be changed), as long as its current pose information can be determined. The current pose information of the camera can be set by the server, and the movement of the camera can be controlled based on the pose information, or the movement of the camera can be controlled by the camera itself or other devices, and the current pose information of the camera can be sent to the server. In some embodiments, more than one camera may be included in the system, and more than one visual sign may also be included.
在一个实施例中,可以为现实场景建立一个场景坐标系(其也可称为现实世界坐标系),并且可以基于摄像头在现实场景中的位姿信息来确定摄像头坐标系与场景坐标系之间的变换矩阵,以及基于视觉标志在现实场景中的位姿信息来确定视觉标志坐标系与场景坐标系之间的变换矩阵。在这种情况下,可以将摄像头坐标系或视觉标志坐标系中的坐标转换为场景坐标系中的坐标,而不在摄像头坐标系和视觉标志坐标系之间进行变换,但是可以理解,摄像头和视觉标志之间的相对位姿信息或变换矩阵仍然能够被服务器知悉。因此,在本申请中,在摄像头和视觉标志之间具有相对位姿是指客观上在两者之间存在相对位姿,而并不要求系统预先存储上述两者之间的相对位姿信息或者使用该相对位姿信息。例如,一个实施例中,在系统中可以仅存储摄像头和视觉标志各自在场景坐标系中的位姿信息,并且可以不计算或者使用两者的相对位姿。In one embodiment, a scene coordinate system (which may also be referred to as a real world coordinate system) may be established for the real scene, and the distance between the camera coordinate system and the scene coordinate system may be determined based on the pose information of the camera in the real scene and the transformation matrix between the visual landmark coordinate system and the scene coordinate system is determined based on the pose information of the visual landmark in the real scene. In this case, it is possible to convert the coordinates in the camera coordinate system or the visual landmark coordinate system to the coordinates in the scene coordinate system without transforming between the camera coordinate system and the visual landmark coordinate system, but it is understood that the camera and visual The relative pose information or transformation matrix between the landmarks can still be known by the server. Therefore, in this application, having a relative pose between the camera and the visual sign refers to objectively having a relative pose between the two, and does not require the system to pre-store the relative pose information between the two above or Use this relative pose information. For example, in one embodiment, only the pose information of the camera and the visual marker in the scene coordinate system may be stored in the system, and the relative poses of the two may not be calculated or used.
摄像头可以用于跟踪现实场景中的目标,该目标可以是静止的或移动的,其例如可以是场景中的人员、静止物体、可移动的物体、等等。可以使用摄像头通过现有技术中的各种方法来跟踪现实场景中的人或物体的位置。例如,对于使用单个单目摄像头的情况,可以结合场景信息(例如,场景中的人或物体所处的平面的信息)来确定场景中的目标的位置信息。对于使用双目摄像头的情况,可以根据目标在摄像头视野中的位置以及目标的深度信息,来确定目标的位置信息。对于使用多个摄像头的情况,可 以根据目标在各个摄像头视野中的位置,来确定目标的位置信息。Cameras can be used to track objects in a real scene, which can be stationary or moving, such as people, stationary objects, movable objects, etc. in the scene. A camera can be used to track the position of a person or object in a real scene by various methods in the prior art. For example, in the case of using a single monocular camera, the location information of objects in the scene can be determined in combination with scene information (eg, information on the plane on which a person or object in the scene is located). For the case of using a binocular camera, the position information of the target can be determined according to the position of the target in the field of view of the camera and the depth information of the target. In the case of using multiple cameras, the position information of the target can be determined according to the position of the target in the field of view of each camera.
可以理解,系统中可以具有多个视觉标志或多个摄像头,多个摄像头的视野可以连续,也可以不连续。It can be understood that the system may have multiple visual signs or multiple cameras, and the fields of view of the multiple cameras may be continuous or discontinuous.
图4示出了根据一个实施例的用于获得场景中的设备或其用户的标识信息的方法,该方法可以使用图3所示的系统来实现,并可以包括如下步骤:Fig. 4 shows a method for obtaining identification information of a device in a scene or a user thereof according to an embodiment. The method can be implemented using the system shown in Fig. 3 and can include the following steps:
步骤401:接收由设备发送的信息,该信息中包括设备或其用户的标识信息以及设备的空间位置信息。Step 401: Receive information sent by the device, where the information includes identification information of the device or its user and spatial location information of the device.
设备发送的信息可以是各种信息,例如报警信息、求助信息、服务请求信息、等等。设备或其用户的标识信息可以是任何能够用于标识或识别设备或其用户的信息,例如设备ID信息、设备的电话号码、设备上某个应用程序的账户信息、用户的姓名或昵称、用户的身份信息、用户的账户信息、等等。The information sent by the device may be various information, such as alarm information, help information, service request information, and so on. The identification information of the device or its user can be any information that can be used to identify or identify the device or its user, such as device ID information, the device's phone number, account information for an application on the device, the user's name or nickname, the user's identity information, user account information, etc.
在一个实施例中,用户303可以使用设备304通过扫描场景中部署的视觉标志301来确定设备304的空间位置信息。用户303可以通过设备304向服务器发送信息,该信息中可以包括设备304的空间位置信息,该空间位置信息可以是设备304相对于视觉标志301的空间位置信息或者设备304在场景中的空间位置信息。在一个实施例中,可以使用设备304采集视觉标志301的图像;通过分析所采集的视觉标志301的图像确定视觉标志301的标识信息以及设备304相对于视觉标志301的空间位置信息;通过视觉标志301的标识信息确定视觉标志301在空间中的位置和姿态信息;以及基于视觉标志301在空间中的位置和姿态信息以及设备304相对于视觉标志301的空间位置信息,确定设备304在场景中的空间位置信息。在一个实施例中,设备304可以向服务器发送视觉标志301的标识信息以及设备304相对于视觉标志301的空间位置信息,从而服务器可以确定设备304在场景中的空间位置信息。In one embodiment, the user 303 may use the device 304 to determine the spatial location information of the device 304 by scanning the visual markers 301 deployed in the scene. The user 303 may send information to the server through the device 304, the information may include the spatial position information of the device 304, and the spatial position information may be the spatial position information of the device 304 relative to the visual sign 301 or the spatial position information of the device 304 in the scene . In one embodiment, an image of the visual sign 301 may be collected using the device 304; identification information of the visual sign 301 and spatial position information of the device 304 relative to the visual sign 301 are determined by analyzing the collected image of the visual sign 301; The identification information of 301 determines the position and attitude information of the visual marker 301 in space; and based on the position and attitude information of the visual marker 301 in space and the spatial position information of the device 304 relative to the visual marker 301, determine the position of the device 304 in the scene. Spatial location information. In one embodiment, the device 304 can send the identification information of the visual marker 301 and the spatial position information of the device 304 relative to the visual marker 301 to the server, so that the server can determine the spatial position information of the device 304 in the scene.
在一个实施例中,还可以使用设备304通过扫描视觉标志301来确定设备304相对于视觉标志301的姿态信息或者设备304在场景中的姿态信息,并可以将该姿态信息发送给服务器。In one embodiment, the device 304 can also be used to scan the visual marker 301 to determine the gesture information of the device 304 relative to the visual marker 301 or the gesture information of the device 304 in the scene, and the gesture information can be sent to the server.
在一个实施例中,设备的空间位置信息和姿态信息可以是设备在扫描 视觉标志时的空间位置信息和姿态信息,也可以是在扫描视觉标志之后的任一时刻的实时位置信息和姿态信息。例如,设备可以在扫描视觉标志时确定其初始的空间位置信息和姿态信息,并继而使用设备内置的各种传感器(例如,加速度传感器、磁力传感器、方向传感器、重力传感器、陀螺仪、摄像头等)通过本领域已知的方法(例如,惯性导航、视觉里程计、SLAM、VSLAM、SFM等)来测量或跟踪其位置变化和/或姿态变化,从而确定设备的实时位置和/或姿态。In one embodiment, the spatial position information and attitude information of the device may be the spatial position information and attitude information of the device when scanning the visual sign, or the real-time position information and attitude information at any moment after scanning the visual sign. For example, a device can determine its initial spatial position information and attitude information when scanning a visual sign, and then use various sensors built into the device (eg, acceleration sensor, magnetic sensor, orientation sensor, gravity sensor, gyroscope, camera, etc.) The real-time position and/or attitude of the device is determined by measuring or tracking its position change and/or attitude change by methods known in the art (eg, inertial navigation, visual odometry, SLAM, VSLAM, SFM, etc.).
服务器所接收的设备的空间位置信息可以是坐标信息,但并不局限于此,任何能够用于推导出设备空间位置的信息都属于空间位置信息。在一个实施例中,服务器所接收的设备的空间位置信息可以是由设备拍摄的视觉标志的图像,服务器可以根据该图像确定设备的空间位置。类似地,任何能够用于推导出设备姿态的信息都属于姿态信息,其在一个实施例中可以是由设备拍摄的视觉标志的图像。The spatial location information of the device received by the server may be coordinate information, but is not limited thereto, any information that can be used to derive the spatial location of the device belongs to spatial location information. In one embodiment, the spatial location information of the device received by the server may be an image of a visual sign captured by the device, and the server may determine the spatial location of the device according to the image. Similarly, any information that can be used to derive a device's pose is pose information, which in one embodiment may be an image of a visual landmark captured by the device.
步骤402:基于设备的空间位置信息识别出在摄像头所拍摄图像中的所述设备或其用户。Step 402: Identify the device or its user in the image captured by the camera based on the spatial location information of the device.
可以使用各种可行的方式通过设备的空间位置信息来在摄像头所拍摄图像中识别出所述设备或其用户。Various feasible ways can be used to identify the device or its user in the image captured by the camera through the spatial location information of the device.
在一个实施例中,可以基于设备的空间位置信息确定设备或其用户在摄像头所拍摄图像中的成像位置,并根据所述成像位置识别出在摄像头所拍摄图像中的所述设备或其用户。In one embodiment, the imaging position of the device or its user in the image captured by the camera may be determined based on the spatial position information of the device, and the device or the user in the image captured by the camera may be identified according to the imaging position.
对于通常由用户持有或携带的设备,例如手机、智能眼镜、智能手表、平板电脑等,可以基于设备的空间位置信息确定其用户在摄像头所拍摄图像中的成像位置。由于用户通常是在手持设备或佩戴设备的状态下扫描视觉标志,因此,可以根据设备的空间位置来推断用户的空间位置,然后根据用户的空间位置确定其在摄像头所拍摄图像中的成像位置。也可以根据设备的空间位置确定其在摄像头所拍摄图像中的成像位置,然后根据设备的成像位置来推断用户的成像位置。For devices usually held or carried by users, such as mobile phones, smart glasses, smart watches, tablet computers, etc., the imaging position of the user in the image captured by the camera can be determined based on the spatial location information of the device. Since the user usually scans the visual sign while holding the device or wearing the device, the spatial position of the user can be inferred according to the spatial position of the device, and then the imaging position of the user in the image captured by the camera can be determined according to the spatial position of the user. The imaging position of the device in the image captured by the camera can also be determined according to the spatial position of the device, and then the imaging position of the user can be inferred according to the imaging position of the device.
对于通常并非由用户持有或携带的设备,例如汽车、机器人、无人驾驶汽车、无人机等,可以基于设备的空间位置信息来确定其在摄像头所拍摄图像中的成像位置。For devices that are not usually held or carried by users, such as automobiles, robots, driverless cars, drones, etc., the imaging position of the device in the image captured by the camera can be determined based on the spatial location information of the device.
在一个实施例中,可以使用预先建立的场景中的一个或多个空间位置(不一定是全部)与摄像头所拍摄图像中的一个或多个成像位置之间的映射关系以及所述设备的空间位置信息,来确定设备或其用户在摄像头所拍摄图像中的成像位置。例如,对于一个大厅场景,可以选择大厅地面上的若干个空间位置,并确定这些位置在摄像头所拍摄图像中的成像位置,之后,可以建立这些空间位置与成像位置之间的映射关系,并可以基于该映射关系推断出与某个空间位置对应的成像位置。In one embodiment, a pre-established mapping relationship between one or more spatial positions (not necessarily all) in the scene and one or more imaging positions in the image captured by the camera and the space of the device may be used Location information to determine the imaging position of the device or its user in the image captured by the camera. For example, for a hall scene, several spatial positions on the floor of the hall can be selected, and the imaging positions of these positions in the image captured by the camera can be determined. After that, the mapping relationship between these spatial positions and the imaging positions can be established, and the An imaging position corresponding to a certain spatial position is deduced based on the mapping relationship.
在一个实施例中,可以基于设备的空间位置信息以及摄像头的位姿信息来确定设备或其用户在摄像头所拍摄图像中的成像位置,其中,摄像头的位姿信息可以是其在场景中的位姿信息或者其相对于视觉标志的位姿信息。In one embodiment, the imaging position of the device or its user in the image captured by the camera may be determined based on the spatial position information of the device and the pose information of the camera, where the pose information of the camera may be its position in the scene pose information or its pose information relative to visual landmarks.
在确定了设备或其用户在摄像头所拍摄图像中的成像位置之后,可以根据该成像位置在所述图像中识别出所述设备或其用户。例如,可以选择与所述成像位置最接近的设备或用户,或者选择与所述成像位置的距离满足预定条件的设备或用户。After the imaging position of the device or its user in the image captured by the camera is determined, the device or its user can be identified in the image according to the imaging position. For example, a device or user closest to the imaging position may be selected, or a device or user whose distance from the imaging position satisfies a predetermined condition may be selected.
在一个实施例中,为了识别出在摄像头所拍摄图像中的所述设备或其用户,可以将设备的空间位置信息与根据摄像头的跟踪结果确定的一个或多个设备或用户的空间位置信息进行比较。可以使用摄像头通过现有技术中的各种方法来确定现实场景中的人或物体的空间位置。例如,对于使用单个单目摄像头的情况,可以结合场景信息(例如,场景中的人或物体所处的平面的信息)来确定场景中的目标的位置信息。对于使用双目摄像头的情况,可以根据目标在摄像头视野中的位置以及目标的深度信息,来确定目标的位置信息。对于使用多个摄像头的情况,可以根据目标在各个摄像头视野中的位置,来确定目标的位置信息。在一个实施例中,还可以使用摄像头拍摄的图像并结合激光雷达等来确定一个或多个用户的空间位置信息。In one embodiment, in order to identify the device or its user in the image captured by the camera, the spatial location information of the device may be compared with the spatial location information of one or more devices or users determined according to the tracking result of the camera. Compare. A camera can be used to determine the spatial position of a person or object in a real scene through various methods in the prior art. For example, in the case of using a single monocular camera, the location information of objects in the scene can be determined in combination with scene information (eg, information on the plane on which a person or object in the scene is located). For the case of using a binocular camera, the position information of the target can be determined according to the position of the target in the field of view of the camera and the depth information of the target. In the case of using multiple cameras, the position information of the target can be determined according to the position of the target in the field of view of each camera. In one embodiment, the spatial location information of one or more users may also be determined by using images captured by a camera in combination with lidar and the like.
在一个实施例中,如果在设备的空间位置附近有多个用户或设备,则可以从所述设备接收其实时的空间位置信息(例如卫星定位信息或者通过设备的传感器获得的位置信息),通过摄像头跟踪所述多个用户或设备的位置,并通过比较从所述设备接收的实时空间位置信息与通过摄像头跟踪 的所述多个用户或设备的位置,来识别所述设备或其用户。In one embodiment, if there are multiple users or devices in the vicinity of the device's spatial location, real-time spatial location information (eg, satellite positioning information or location information obtained through the device's sensors) may be received from the device by A camera tracks the location of the plurality of users or devices and identifies the device or a user thereof by comparing real-time spatial location information received from the device to the locations of the plurality of users or devices tracked by the camera.
在一个实施例中,如果在设备的空间位置附近有多个用户,则可以基于设备发送的信息确定设备用户的特征信息(例如,用于人脸识别的特征信息),通过摄像头采集所述多个用户的特征信息,并通过比较所述多个用户的特征信息与设备用户的特征信息,来识别设备用户。In one embodiment, if there are multiple users near the spatial location of the device, feature information of the device user (for example, feature information for face recognition) may be determined based on information sent by the device, and the multiple users may be collected by a camera. feature information of a plurality of users, and identify the device user by comparing the feature information of the plurality of users with the feature information of the device user.
在一个实施例中,可以基于设备的空间位置信息首先确定视野能够覆盖设备或其用户的一个或多个摄像头,并继而确定设备或其用户在该一个或多个摄像头所拍摄图像中的成像位置。In one embodiment, based on the spatial location information of the device, it is firstly determined that the field of view can cover one or more cameras of the device or its user, and then the imaging position of the device or its user in the images captured by the one or more cameras is determined .
步骤403:将所述设备或其用户的标识信息关联到在摄像头所拍摄图像中的所述设备或其用户,以便使用所述标识信息向所述设备或其用户提供服务。Step 403: Associate the identification information of the device or its user with the device or its user in the image captured by the camera, so as to use the identification information to provide a service to the device or its user.
在识别出在摄像头所拍摄图像中的所述设备或其用户之后,可以将所接收的设备或其用户的标识信息关联到图像中的所述设备或其用户。如此,例如可以知悉摄像头视野中的设备的ID信息、电话号码、设备上某个应用程序的账户信息,或者可以知悉摄像头视野中的用户的姓名或昵称、用户的身份信息、用户的账户信息、等等。在知悉了摄像头视野中的设备或用户的标识信息之后,可以使用该标识信息向所述设备或其用户提供各种服务,例如,导航服务、讲解服务、信息展示服务、等等。在一个实施例中,可以以视觉、声音等方式来提供上述信息。在一个实施例中,可以在设备(例如手机或眼镜)的显示媒介上叠加虚拟对象,该虚拟对象例如可以是图标(例如导航图标)、图片、文字等。After identifying the device or its user in the image captured by the camera, the received identification information of the device or its user may be associated with the device or its user in the image. In this way, for example, the ID information, phone number, account information of an application on the device can be known, or the user's name or nickname, the user's identity information, the user's account information, and many more. After knowing the identification information of the device or user in the field of view of the camera, the identification information can be used to provide various services to the device or its user, such as navigation service, explanation service, information display service, and so on. In one embodiment, the above information may be provided visually, audibly, or the like. In one embodiment, a virtual object may be superimposed on a display medium of a device (eg, a mobile phone or glasses), and the virtual object may be, for example, an icon (eg, a navigation icon), a picture, a text, and the like.
图4所示的方法中的步骤可以由图3所示的系统中的服务器来实现,但可以理解,这些步骤中的一个或多个也可以由其他装置来实现。The steps in the method shown in FIG. 4 may be implemented by the server in the system shown in FIG. 3 , but it is understood that one or more of these steps may also be implemented by other devices.
在一个实施例中,还可以通过摄像头跟踪场景中所述设备或其用户以获得其实时的位置信息和/或姿态信息,或者可以使用所述设备来获得其实时的位置信息和/或姿态信息。在获得了设备或其用户的位置和/或姿态信息之后,可以基于该位置和/或姿态信息向设备或其用户提供服务。In one embodiment, the device or its user in the scene can also be tracked through a camera to obtain its real-time position information and/or attitude information, or the device can be used to obtain its real-time position information and/or attitude information . After the location and/or attitude information of the device or its user is obtained, services can be provided to the device or its user based on the location and/or attitude information.
在一个实施例中,在将所述设备或其用户的标识信息关联到在摄像头所拍摄图像中的所述设备或其用户后,可以通过该标识信息向摄像头视野中的相应设备或用户发送信息,该信息例如是导航信息、讲解信息、指示 信息、广告信息、等等。In one embodiment, after the identification information of the device or its user is associated with the device or its user in the image captured by the camera, information can be sent to the corresponding device or user in the field of view of the camera through the identification information , the information is, for example, navigation information, explanation information, instruction information, advertisement information, and so on.
下文描述本文的一个具体应用场景。A specific application scenario of this paper is described below.
在一个使用机器人运送货物的智能化工厂场景中部署了一个或多个视觉标志以及一个或多个摄像头。在机器人行进过程中,使用摄像头对机器人的位置进行跟踪,并根据所跟踪的位置向机器人发送导航指示。为了确定摄像头视野中的每个机器人的标识信息(例如,机器人的ID),可以使得每个机器人例如在进入场景或者摄像头视野范围时扫描视觉标志,并发送其位置信息和标识信息。如此,可以容易地确定摄像头视野范围内的每个机器人的标识信息,从而基于每个机器人的当前位置以及其要完成的工作任务向其发送行进指示或导航指示。One or more visual signs and one or more cameras are deployed in a smart factory scenario where robots are used to deliver goods. During the movement of the robot, the camera is used to track the position of the robot, and navigation instructions are sent to the robot according to the tracked position. In order to determine the identification information (eg, robot ID) of each robot in the camera's field of view, each robot may be made to scan a visual sign, for example, when entering the scene or the camera's field of view, and send its position information and identification information. In this way, the identification information of each robot within the field of view of the camera can be easily determined, so as to send each robot a travel instruction or a navigation instruction based on its current position and the work task to be completed.
在一个实施例中,可以向设备发送与虚拟对象有关的信息,该虚拟对象例如可以是图片、文字、数字、图标、视频、三维模型等,与虚拟对象有关的信息可以包括虚拟对象的空间位置信息。在设备接收到虚拟对象后,可以在设备的显示媒介上呈现该虚拟对象。在一个实施例中,设备可以基于设备或用户的空间位置信息和/或姿态信息,在其显示媒介上的适当位置处呈现该虚拟对象。该虚拟对象例如可以以增强现实或混合现实的方式呈现在用户设备的显示媒介上。在一个实施例中,所述虚拟对象是通过对直播人物进行视频采集而生成的视频影像或动态三维模型。例如,该虚拟对象可以是对服务人员进行实时视频采集而生成的视频影像,该视频影像可以呈现于用户设备的显示媒介上,从而向用户提供服务。在一个实施例中,可以设置上述视频影像的空间位置,使得其可以以增强现实或混合现实的方式呈现在用户设备的显示媒介上。In one embodiment, information related to a virtual object may be sent to the device, the virtual object may be, for example, pictures, characters, numbers, icons, videos, three-dimensional models, etc., and the information related to the virtual object may include the spatial location of the virtual object information. After the device receives the virtual object, the virtual object can be presented on the display medium of the device. In one embodiment, the device may present the virtual object at an appropriate location on its display medium based on the device's or user's spatial location information and/or gesture information. The virtual object may be presented on the display medium of the user equipment in an augmented reality or mixed reality manner, for example. In one embodiment, the virtual object is a video image or a dynamic three-dimensional model generated by video capture of live characters. For example, the virtual object may be a video image generated by real-time video capture of service personnel, and the video image may be presented on the display medium of the user equipment, so as to provide services to the user. In one embodiment, the spatial position of the video image can be set so that it can be presented on the display medium of the user equipment in the manner of augmented reality or mixed reality.
在一个实施例中,在将所述设备或其用户的标识信息关联到在摄像头所拍摄图像中的所述设备或其用户后,可以基于该标识信息识别由摄像头视野内的设备或用户发送的信息,该信息例如是服务请求信息、报警信息、求助信息、评论信息等。在一个实施例中,在接收到所述设备或用户发送的信息之后,可以根据该信息设置与所述设备或用户相关联的虚拟对象,其中,该虚拟对象的空间位置信息可以根据设备或用户的位置信息确定,并且可以随着设备或用户的位置变化而相应地改变虚拟对象的空间位置。如此,其他用户可以通过一些设备(例如,手机、智能眼镜等)通过增强 现实或混合现实的方式观察到上述虚拟对象。在一个实施例中,可以根据从所述设备或用户接收到的新的信息(例如,用户的新的评论)更新虚拟对象的内容(例如,更新虚拟对象的文字内容)。In one embodiment, after the identification information of the device or its user is associated with the device or its user in the image captured by the camera, the identification information sent by the device or user within the field of view of the camera can be identified based on the identification information. information, such as service request information, alarm information, help information, comment information, and the like. In one embodiment, after receiving the information sent by the device or the user, a virtual object associated with the device or the user may be set according to the information, wherein the spatial location information of the virtual object may be based on the information of the device or the user The position information of the virtual object can be determined, and the spatial position of the virtual object can be changed accordingly as the position of the device or the user changes. In this way, other users can observe the above-mentioned virtual objects by means of augmented reality or mixed reality through some devices (for example, mobile phones, smart glasses, etc.). In one embodiment, the content of the virtual object (eg, updating the textual content of the virtual object) may be updated according to new information received from the device or user (eg, a new comment by the user).
图5示出了根据一个实施例的用于向场景中的设备或用户提供服务的方法,该方法可以使用图3所示的系统来实现,并可以包括如下步骤:Fig. 5 shows a method for providing a service to a device or a user in a scene according to one embodiment. The method can be implemented using the system shown in Fig. 3 and can include the following steps:
步骤501:接收由设备发送的信息,该信息中包括设备或其用户的标识信息以及设备的空间位置信息。Step 501: Receive information sent by the device, where the information includes identification information of the device or its user and spatial location information of the device.
步骤502:基于设备的空间位置信息识别出在摄像头所拍摄图像中的所述设备或其用户。Step 502: Identify the device or its user in the image captured by the camera based on the spatial location information of the device.
步骤503:在摄像头所拍摄图像中标示出所述设备或其用户。Step 503: Mark the device or its user in the image captured by the camera.
可以使用多种方法来标示设备或用户,例如,可以框出设备或用户的成像,可以将特定的图标呈现于设备或用户的成像附近,或者可以高亮显示设备或用户的成像。在一个实施例中,可以放大所标示的设备或用户的成像区域,或者使摄像头针对所标示的设备或用户进行拍摄。在一个实施例中,可以通过摄像头持续跟踪所述设备或用户,并且可以确定设备或用户的实时空间位置信息和/或姿态信息。The device or user can be identified using a variety of methods, for example, an image of the device or user can be framed, a particular icon can be presented adjacent to the device or user's image, or the device or user's image can be highlighted. In one embodiment, the imaging area of the marked device or user can be enlarged, or the camera can be made to shoot for the marked device or user. In one embodiment, the device or user can be continuously tracked through a camera, and real-time spatial location information and/or gesture information of the device or user can be determined.
步骤504:将所述设备或其用户的标识信息关联到在摄像头所拍摄图像中的所述设备或其用户,以便使用所述标识信息向所述设备或其用户提供服务。Step 504: Associate the identification information of the device or its user with the device or its user in the image captured by the camera, so as to use the identification information to provide services to the device or its user.
在摄像头所拍摄图像中标示出设备或用户之后,能够观察摄像头所拍摄图像的人员(例如,机场、车站、商场的管理人员或服务人员)可以知道该设备或用户当前需要服务,并可以知道该设备或用户目前所处的位置,从而可以方便地向该设备或用户提供各种需要的服务,例如讲解服务、导航服务、咨询服务、帮助服务等。通过这种方式,可以替代场景中部署的咨询台,并可以以便捷、低成本的方式向场景中的任意用户提供其需要的服务。After a device or user is marked in the image captured by the camera, a person who can observe the image captured by the camera (for example, management or service personnel in airports, stations, shopping malls) can know that the device or user currently needs service, and can know that the device or user currently needs service. The current location of the device or user, so that various required services, such as explanation service, navigation service, consulting service, help service, etc., can be conveniently provided to the device or user. In this way, the help desk deployed in the scenario can be replaced, and any user in the scenario can be provided with the services they need in a convenient and low-cost manner.
在一个实施例中,可以通过用户携带或控制的设备来向用户提供服务,该设备例如可以是手机、智能眼镜、车辆等。在一个实施例中,可以通过设备上的电话功能、应用程序(APP)等以视觉、声音等方式来提供服务。In one embodiment, the service may be provided to the user through a device carried or controlled by the user, such as a mobile phone, smart glasses, a vehicle, and the like. In one embodiment, the service may be provided visually, audibly, etc. through a telephony function, an application (APP), etc. on the device.
图5所示的方法中的步骤可以由图3所示的系统中的服务器来实现,但可以理解,这些步骤中的一个或多个也可以由其他装置来实现。The steps in the method shown in FIG. 5 may be implemented by the server in the system shown in FIG. 3 , but it is understood that one or more of these steps may also be implemented by other devices.
图6示出了根据一个实施例的用于通过设备(在此以眼镜为例)向场景中的用户提供信息的方法,该方法可以使用图3所示的系统来实现,并可以包括如下步骤:Fig. 6 shows a method for providing information to a user in a scene through a device (here, glasses are taken as an example) according to an embodiment, the method can be implemented using the system shown in Fig. 3, and can include the following steps :
步骤601:接收由眼镜发送的信息,该信息中包括眼镜的空间位置信息。Step 601: Receive information sent by the glasses, where the information includes spatial position information of the glasses.
在一个实施例中,用户可以使用眼镜通过扫描场景中部署的视觉标志来确定眼镜的空间位置信息。用户可以通过眼镜向服务器发送信息。在一个实施例中,还可以使用眼镜通过扫描视觉标志来确定眼镜相对于视觉标志的姿态信息或者眼镜在场景中的姿态信息,并可以将该姿态信息发送给服务器。In one embodiment, the user may use the glasses to determine the spatial position information of the glasses by scanning the visual landmarks deployed in the scene. The user can send information to the server through the glasses. In one embodiment, the glasses can also be used to scan the visual markers to determine the gesture information of the glasses relative to the visual markers or the gesture information of the glasses in the scene, and the gesture information can be sent to the server.
在一个实施例中,眼镜发送的信息中除了包括眼镜的空间位置信息之外,还可以包括眼镜或其用户的相关信息,例如服务请求信息、求助信息、报警信息、标识信息(例如电话号码、APP账户信息)等。In one embodiment, in addition to the spatial location information of the glasses, the information sent by the glasses may also include information related to the glasses or their users, such as service request information, help information, alarm information, identification information (such as phone numbers, APP account information) etc.
在一个实施例中,眼镜本身可以具备直接接入网络的能力。在另一个实施例中,眼镜可以不具备直接接入网络的能力,而是通过其与例如用户手机之间的连接来间接地接入网络,在这种情况下,服务器可以通过手机等中间设备接收由眼镜发送的信息。In one embodiment, the glasses themselves may be capable of direct access to the network. In another embodiment, the glasses may not have the ability to directly access the network, but indirectly access the network through a connection between it and, for example, the user's mobile phone. In this case, the server may use an intermediate device such as a mobile phone. Receive information sent by glasses.
步骤602:基于眼镜的空间位置信息识别出在摄像头所拍摄图像中的所述眼镜的用户。Step 602: Identify the user of the glasses in the image captured by the camera based on the spatial position information of the glasses.
如上文所述,可以使用各种可行的方式通过眼镜的空间位置信息来在摄像头所拍摄图像中识别出眼镜的用户。As described above, various feasible ways can be used to identify the user of the glasses in the image captured by the camera through the spatial position information of the glasses.
在识别出用户之后,可以将用户的标识信息关联到用户,以便使用所述标识信息向所述用户提供服务。After the user is identified, the user's identification information can be associated with the user in order to provide services to the user using the identification information.
步骤603:通过摄像头跟踪用户并更新用户的空间位置信息。Step 603: Track the user through the camera and update the spatial location information of the user.
在一个实施例中,可以通过摄像头跟踪用户并更新所述用户的成像位置,并基于所述更新的成像位置确定用户的空间位置信息。可以使用本领域已知的各种视觉跟踪方法来跟踪摄像头视野中的用户并更新所述用户的成像位置。在跟踪用户的过程中,摄像头可以保持固定不动或者可移动。 在一个实施例中,在跟踪用户的过程中,可以使用多个摄像头,这些摄像头可以具有连续的视野范围或者不连续的视野范围。在视野范围不连续的情况下,可以记录用户的特征,并当用户再次进入一个或多个摄像头的视野范围时重新识别用户并跟踪。In one embodiment, a camera may be used to track the user and update the imaging position of the user, and determine the spatial position information of the user based on the updated imaging position. Various visual tracking methods known in the art can be used to track the user in the field of view of the camera and update the imaging position of the user. The camera can remain stationary or move while tracking the user. In one embodiment, in tracking the user, multiple cameras may be used, which may have a continuous field of view or a discontinuous field of view. Where the field of view is discontinuous, the user's characteristics can be recorded and re-identified and tracked when the user re-enters the field of view of one or more cameras.
在一个实施例中,可以使用预先建立的场景中的一个或多个空间位置(不一定是全部)与摄像头所拍摄图像中的一个或多个成像位置之间的映射关系以及所述成像位置,来确定用户的空间位置信息。在一个实施例中,可以基于摄像头的位姿信息和所述成像位置来确定用户的空间位置信息。例如,在使用深度摄像头或者多目摄像头的情况下,可以基于所述成像位置确定用户相对于摄像头的方向,可以使用深度信息确定用户相对于摄像头的距离,从而确定用户相对于摄像头的位置,继而,可以进一步基于摄像头的位姿信息确定用户的空间位置信息。在一个实施例中,可以基于用户的成像大小来估计用户相对于摄像头的距离,并基于摄像头的位姿信息和所述成像位置来确定用户的空间位置信息。在一个实施例中,可以使用摄像头上安装的激光雷达等来确定用户相对于摄像头的距离,并基于摄像头的位姿信息和所述成像位置来确定用户的空间位置信息。在一个实施例中,如果多个摄像头的视野同时覆盖用户,可以使用该多个摄像头来共同确定用户的空间位置信息。在一个实施例中,可以基于摄像头的位姿信息、所述成像位置以及可选的其他信息(例如场景内地面的坐标信息)来确定用户的空间位置信息。In one embodiment, a pre-established mapping relationship between one or more spatial positions (not necessarily all) in the scene and one or more imaging positions in the image captured by the camera and the imaging positions may be used, to determine the user's spatial location information. In one embodiment, the spatial position information of the user may be determined based on the pose information of the camera and the imaging position. For example, in the case of using a depth camera or a multi-camera camera, the direction of the user relative to the camera can be determined based on the imaging position, the depth information can be used to determine the distance of the user relative to the camera, so as to determine the position of the user relative to the camera, and then , the spatial position information of the user can be further determined based on the pose information of the camera. In one embodiment, the distance of the user relative to the camera may be estimated based on the imaging size of the user, and the spatial position information of the user may be determined based on the pose information of the camera and the imaging position. In one embodiment, the distance of the user relative to the camera may be determined by using a lidar or the like installed on the camera, and the spatial position information of the user may be determined based on the pose information of the camera and the imaging position. In one embodiment, if the visual fields of multiple cameras cover the user at the same time, the multiple cameras can be used to jointly determine the spatial location information of the user. In one embodiment, the spatial position information of the user may be determined based on the pose information of the camera, the imaging position, and optional other information (eg, coordinate information of the ground in the scene).
在一个实施例中,还可以基于摄像头对用户的跟踪结果来确定用户的姿态信息。In one embodiment, the user's gesture information may also be determined based on the tracking result of the user by the camera.
步骤604:基于用户的空间位置信息通过用户的眼镜向用户提供信息。Step 604: Provide information to the user through the user's glasses based on the user's spatial location information.
在知悉用户的空间位置信息的情况下,可以向用户提供各种需要的信息,例如,导航信息、指示信息、讲解信息、广告信息、与基于位置的服务有关的其他信息、等等。在一个实施例中,可以以视觉、声音等方式来提供上述信息。在一个实施例中,可以在眼镜的显示媒介上叠加虚拟对象,该虚拟对象例如可以是图标(例如导航图标)、图片、文字等。Knowing the user's spatial location information, the user can be provided with various required information, such as navigation information, instruction information, tutorial information, advertising information, other information related to location-based services, and the like. In one embodiment, the above information may be provided visually, audibly, or the like. In one embodiment, a virtual object may be superimposed on the display medium of the glasses, and the virtual object may be, for example, an icon (eg, a navigation icon), a picture, a text, or the like.
在一个实施例中,眼镜本身可以具备直接接入网络的能力,从而眼镜可以直接从服务器接收指示信息。在另一个实施例中,眼镜可以不具备直 接接入网络的能力,而是通过其与例如用户手机之间的连接来间接地接入网络,在这种情况下,眼镜可以通过手机等中间设备从服务器接收指示信息。In one embodiment, the glasses themselves may have the ability to directly access the network, so that the glasses may directly receive indication information from the server. In another embodiment, the glasses may not have the ability to directly access the network, but indirectly access the network through a connection between it and, for example, the user's mobile phone, in this case, the glasses may pass through an intermediate device such as a mobile phone Receive instructions from the server.
在一个实施例中,可以进一步结合眼镜或其用户的姿态信息来向用户提供信息。可以通过眼镜确定眼镜或其用户的姿态信息,或者通过摄像头所拍摄的用户图像来确定用户的姿态信息,该姿态信息可以包括用户的朝向信息。在一个实施例中,眼镜的姿态信息可以通过其内置的传感器来获得,例如,通过对初始姿态进行跟踪的方式或者直接由眼镜的内置传感器(例如,重力传感器、磁力传感器、方向传感器等)确定。服务器可以直接从眼镜接收所述姿态信息,或者通过手机等中间设备接收所述姿态信息。In one embodiment, information may be further provided to the user in conjunction with the glasses or the gesture information of the user thereof. The posture information of the glasses or the user thereof may be determined by the glasses, or the posture information of the user may be determined by the user image captured by the camera, and the posture information may include the orientation information of the user. In one embodiment, the posture information of the glasses can be obtained through its built-in sensors, for example, by tracking the initial posture or directly determined by the built-in sensors of the glasses (for example, a gravity sensor, a magnetic sensor, an orientation sensor, etc.) . The server may directly receive the gesture information from the glasses, or receive the gesture information through an intermediate device such as a mobile phone.
图6所示的方法中的步骤可以由图3所示的系统中的服务器来实现,但可以理解,这些步骤中的一个或多个也可以由其他装置来实现。The steps in the method shown in FIG. 6 may be implemented by the server in the system shown in FIG. 3 , but it is understood that one or more of these steps may also be implemented by other devices.
图7示出了根据一个实施例的通过眼镜向场景中的用户提供信息的系统,该系统包括视觉标志701、摄像头702以及服务器(图7中未示出)。用户703位于场景中,并且携带有眼镜704和手机705。手机705可以通过其上的图像采集器件来识别视觉标志701,如此,眼镜704上可以不具有图像采集器件,或者其上虽然具有图像采集器件该该图像采集器件可以不具有识别视觉标志701的能力。FIG. 7 shows a system for providing information to a user in a scene through glasses, including a visual sign 701, a camera 702, and a server (not shown in FIG. 7), according to one embodiment. A user 703 is in the scene and carries glasses 704 and a mobile phone 705 . The mobile phone 705 can recognize the visual sign 701 through the image capture device on it, so the glasses 704 may not have an image capture device, or although there is an image capture device on it, the image capture device may not have the ability to recognize the visual sign 701 .
图8示出了根据一个实施例的通过眼镜向场景中的用户提供信息的方法,该方法可以使用图7所示的系统来实现。该方法包括如下步骤(部分步骤与图6中的步骤类似,在此不再赘述,但可以理解,针对图6中各个步骤描述的内容同样可以适用于图8中的相应步骤):FIG. 8 illustrates a method of providing information to a user in a scene through glasses, which may be implemented using the system shown in FIG. 7 , according to one embodiment. The method includes the following steps (part of the steps are similar to the steps in FIG. 6 , and will not be repeated here, but it can be understood that the content described for each step in FIG. 6 can also be applied to the corresponding steps in FIG. 8 ):
步骤801:接收由用户的手机发送的信息,该信息中包括手机的空间位置信息。Step 801: Receive information sent by the user's mobile phone, where the information includes spatial location information of the mobile phone.
用户可以使用手机通过扫描场景中部署的视觉标志来确定手机的空间位置信息。在一个实施例中,还可以通过扫描视觉标志来确定手机的姿态信息,并可以将该姿态信息发送给服务器。The user can use the mobile phone to determine the spatial location information of the mobile phone by scanning the visual landmarks deployed in the scene. In one embodiment, the gesture information of the mobile phone can also be determined by scanning the visual sign, and the gesture information can be sent to the server.
步骤802:基于手机的空间位置信息识别出在摄像头所拍摄图像中的所述手机的用户。Step 802: Identify the user of the mobile phone in the image captured by the camera based on the spatial location information of the mobile phone.
在识别出用户之后,可以将用户的标识信息关联到用户,以便使用所述标识信息向所述用户提供服务。After the user is identified, the user's identification information can be associated with the user in order to provide services to the user using the identification information.
步骤803:通过摄像头跟踪用户并更新用户的空间位置信息。Step 803: Track the user through the camera and update the spatial location information of the user.
在一个实施例中,还可以确定用户的姿态信息。In one embodiment, the user's gesture information can also be determined.
步骤804:基于用户的空间位置信息通过用户的眼镜向用户提供信息。Step 804: Provide information to the user through the user's glasses based on the user's spatial location information.
在一个实施例中,眼镜本身可以具备直接接入网络的能力,从而眼镜可以直接从服务器接收指示信息。在另一个实施例中,眼镜可以不具备直接接入网络的能力,而是通过其与例如用户手机之间的连接来间接地接入网络,在这种情况下,眼镜可以通过手机等中间设备从服务器接收指示信息。例如,服务器可以首先向用户的手机发送第一信息,之后手机可以基于所述第一信息向眼镜发送第二信息(第二信息可以与第一信息相同或不同),以通过眼镜向用户提供基于位置的服务。In one embodiment, the glasses themselves may have the ability to directly access the network, so that the glasses may directly receive indication information from the server. In another embodiment, the glasses may not have the ability to directly access the network, but indirectly access the network through a connection between it and, for example, the user's mobile phone, in this case, the glasses may pass through an intermediate device such as a mobile phone Receive instructions from the server. For example, the server may first send first information to the user's mobile phone, and then the mobile phone may send second information (the second information may be the same as or different from the first information) to the glasses based on the first information, so as to provide the user with information based on the glasses through the glasses. location services.
在一个实施例中,可以进一步结合眼镜或其用户的姿态信息来向用户提供信息。In one embodiment, information may be further provided to the user in conjunction with the glasses or the gesture information of the user thereof.
在一个实施例中,用户也可以不使用眼镜,而是仅使用手机。如此,在上述步骤804中,可以基于用户的空间位置信息通过用户的手机向用户提供信息。在一个实施例中,可以进一步结合手机或其用户的姿态信息来向用户提供信息。可以通过手机确定其姿态信息,或者通过摄像头所拍摄的用户图像来确定用户的姿态信息。在一个实施例中,手机的姿态信息可以通过其内置传感器来获得。In one embodiment, the user may also not use the glasses, but only use the cell phone. In this way, in the above step 804, information may be provided to the user through the user's mobile phone based on the user's spatial location information. In one embodiment, the information may be further provided to the user in combination with the gesture information of the mobile phone or its user. The gesture information of the user can be determined through the mobile phone, or the user's gesture information can be determined through the user image captured by the camera. In one embodiment, the gesture information of the mobile phone can be obtained through its built-in sensor.
在本申请中,用于扫描视觉标志来确定其空间位置信息的设备可以被称为“位置获得设备”,用于向用户提供信息的设备可以被称为“信息接收设备”。根据本申请上文的描述可以理解,位置获得设备和信息接收设备可以是同一设备,例如用户的手机或者用户的眼镜;位置获得设备和信息接收设备也可以是不同的设备,例如分别是手机和眼镜。In this application, a device for scanning a visual sign to determine its spatial location information may be referred to as a "position acquisition device", and a device for providing information to a user may be referred to as an "information receiving device". It can be understood from the above description of this application that the location obtaining device and the information receiving device may be the same device, such as the user's mobile phone or the user's glasses; the location obtaining device and the information receiving device may also be different devices, such as a mobile phone and a user's glasses respectively. Glasses.
图9示出了根据一个实施例的用户交互系统,该系统包括视觉标志901、摄像头902以及服务器(图9中未示出)。摄像头和视觉标志各自以特定的位置和姿态(下文中可以统称为“位姿”)部署于现实场景中,场景中还具有第一用户903和第二用户905,其分别携带有第一设备904和第二设备906。第一设备904和第二设备906上具有图像采集器件并能够 通过图像采集器件识别视觉标志901。第一设备904和第二设备906例如可以是手机、眼镜等设备。Figure 9 illustrates a user interaction system including a visual sign 901, a camera 902, and a server (not shown in Figure 9) according to one embodiment. The camera and the visual sign are each deployed in a real scene with a specific position and attitude (hereinafter collectively referred to as "pose"), and the scene also has a first user 903 and a second user 905 who carry a first device 904 respectively. and the second device 906. The first device 904 and the second device 906 have image capture devices on them and can identify the visual sign 901 through the image capture devices. The first device 904 and the second device 906 may be, for example, mobile phones, glasses and other devices.
图10示出了根据一个实施例的用户交互方法,其可以使用上述系统来实现,并且可以包括如下步骤:FIG. 10 shows a user interaction method according to one embodiment, which can be implemented using the above-mentioned system, and can include the following steps:
步骤1001:接收由第一用户的第一设备发送的信息,该信息中包括第一设备的空间位置信息以及第一用户或第一设备的标识信息。Step 1001: Receive information sent by a first device of a first user, where the information includes spatial location information of the first device and identification information of the first user or the first device.
在一个实施例中,第一用户可以使用第一设备通过扫描场景中部署的视觉标志来确定第一设备的空间位置信息。在一个实施例中,还可以使用第一设备通过扫描视觉标志来确定第一设备相对于视觉标志的姿态信息或者第一设备在场景中的姿态信息,并可以将该姿态信息发送给服务器。In one embodiment, the first user may use the first device to determine the spatial location information of the first device by scanning the visual markers deployed in the scene. In one embodiment, the first device can also be used to scan the visual marker to determine the gesture information of the first device relative to the visual marker or the gesture information of the first device in the scene, and the gesture information can be sent to the server.
步骤1002:基于第一设备的空间位置信息识别出在摄像头所拍摄图像中的所述第一用户。Step 1002: Identify the first user in the image captured by the camera based on the spatial location information of the first device.
步骤1003:将第一用户或第一设备的标识信息关联到在摄像头所拍摄图像中的所述第一用户。Step 1003: Associate the identification information of the first user or the first device with the first user in the image captured by the camera.
步骤1004:通过摄像头跟踪第一用户并更新第一用户的空间位置信息。Step 1004: Track the first user through the camera and update the spatial location information of the first user.
在一个实施例中,还可以基于摄像头对用户或设备的跟踪结果来确定用户或设备的姿态信息。In one embodiment, the gesture information of the user or the device may also be determined based on the tracking result of the user or the device by the camera.
步骤1005:设置与所述第一用户关联的第一虚拟对象的相关信息,所述相关信息包括内容信息和空间位置信息,其中,根据所述第一用户的空间位置信息设置第一虚拟对象的空间位置信息。Step 1005: Set relevant information of the first virtual object associated with the first user, the relevant information includes content information and spatial position information, wherein the first virtual object is set according to the spatial position information of the first user. Spatial location information.
例如,可以将第一虚拟对象的空间位置配置为位于第一用户上方的预定距离处。第一虚拟对象的内容信息是用于描述该虚拟对象的内容的相关信息,其例如可以包括虚拟对象中包含的图片、文字、数字、图标、动画、视频、三维模型等,也可以包括虚拟对象的形状信息、颜色信息、尺寸信息、姿态信息等。在一个实施例中,可以根据通过第一用户或第一设备的标识信息识别的来自所述第一用户或第一设备的信息设置第一虚拟对象的内容信息。在一个实施例中,第一虚拟对象的内容信息例如可以是第一用户的职业、身份、性别、年龄、姓名、昵称等。For example, the spatial position of the first virtual object may be configured to be at a predetermined distance above the first user. The content information of the first virtual object is related information used to describe the content of the virtual object, which may include, for example, pictures, characters, numbers, icons, animations, videos, three-dimensional models, etc. contained in the virtual object, and may also include virtual objects shape information, color information, size information, posture information, etc. In one embodiment, the content information of the first virtual object may be set according to the information from the first user or the first device identified by the identification information of the first user or the first device. In one embodiment, the content information of the first virtual object may be, for example, the occupation, identity, gender, age, name, nickname, etc. of the first user.
第一虚拟对象的空间位置信息可以随着第一用户的位置变化而相应 地变化,并且可以根据从第一用户或第一设备接收到的新的信息(例如,用户的新的评论)更新虚拟对象的内容信息(例如,更新虚拟对象的文字内容)。The spatial location information of the first virtual object may change accordingly as the location of the first user changes, and the virtual object may be updated according to new information received from the first user or the first device (eg, new comments by the user). Content information of the object (eg, updating the textual content of the virtual object).
在一个实施例中,还可以设置虚拟对象的姿态信息,虚拟对象的姿态信息可以基于与其关联的设备或用户的姿态信息来设置,但也可以通过其他方式来设置。In one embodiment, the pose information of the virtual object may also be set, and the pose information of the virtual object may be set based on the pose information of the device or user associated therewith, but may also be set in other ways.
步骤1006:将所述第一虚拟对象的相关信息发送给第二用户的第二设备。Step 1006: Send the relevant information of the first virtual object to the second device of the second user.
所述第一虚拟对象的相关信息能够被所述第二设备使用以基于其位置信息和/或姿态信息在其显示媒介上呈现所述第一虚拟对象(例如以增强现实或混合现实的方式)。Information about the first virtual object can be used by the second device to render the first virtual object on its display medium based on its position information and/or gesture information (eg, in an augmented or mixed reality manner) .
可以以各种可行的方式来确定第二设备的位置信息和姿态信息。在一个实施例中,第二设备可以通过扫描视觉标志来确定其位置信息和/或姿态信息。在一个实施例中,可以通过摄像头对第二设备或其用户的跟踪结果来确定第二设备的位置信息和/或姿态信息。在一个实施例中,第二设备也可以使用其内置的各种传感器来确定其位置信息和/或姿态信息。在一个实施例中,第二设备可以使用场景的点云信息来确定其位置信息和/或姿态信息。The location information and attitude information of the second device may be determined in various feasible ways. In one embodiment, the second device may determine its position information and/or gesture information by scanning the visual landmarks. In one embodiment, the location information and/or posture information of the second device may be determined through the tracking result of the second device or its user by the camera. In one embodiment, the second device may also use various sensors built in it to determine its position information and/or attitude information. In one embodiment, the second device may use point cloud information of the scene to determine its position information and/or pose information.
在一个实施例中,在获得了第一虚拟对象的空间位置信息以及第二设备的位置和姿态信息之后,便可以在通过第二设备的显示媒介呈现的现实场景中的合适位置处叠加第一虚拟对象。在第一虚拟对象具有姿态信息的情况下,可以进一步确定所叠加的第一虚拟对象的姿态。In one embodiment, after obtaining the spatial position information of the first virtual object and the position and attitude information of the second device, the first virtual object can be superimposed at a suitable position in the real scene presented by the display medium of the second device virtual object. In the case that the first virtual object has gesture information, the gesture of the superimposed first virtual object may be further determined.
在一个实施例中,在叠加了第一虚拟对象之后,第二设备的用户可以对第一虚拟对象执行各种交互操作。In one embodiment, after the first virtual object is superimposed, the user of the second device may perform various interactive operations on the first virtual object.
在一个实施例中,也可以以类似的方式为第二设备的第二用户设置第二虚拟对象,并将该第二虚拟对象的内容信息和空间位置信息发送给第一用户的第一设备或其他设备(该第一设备和其他设备例如可以分别是手机和眼镜),其中,所述第二虚拟对象的内容信息和空间位置信息能够被所述第一设备或其他设备使用以基于其位置信息和/或姿态信息在其显示媒介上呈现所述第二虚拟对象。In one embodiment, a second virtual object may also be set for the second user of the second device in a similar manner, and the content information and spatial location information of the second virtual object may be sent to the first device or the first device of the first user. other devices (the first device and the other devices may be, for example, a mobile phone and glasses, respectively), wherein the content information and the spatial location information of the second virtual object can be used by the first device or the other devices to be based on its location information and/or gesture information to present the second virtual object on its display medium.
图10所示的方法中的步骤可以由图9所示的系统中的服务器来实现,但可以理解,这些步骤中的一个或多个也可以由其他装置来实现。The steps in the method shown in FIG. 10 may be implemented by the server in the system shown in FIG. 9 , but it is understood that one or more of these steps may also be implemented by other devices.
图11示出了根据一个实施例的由第二用户通过其设备(例如眼镜或手机)观察到的第一用户以及与该第一用户关联的虚拟对象。该虚拟对象例如可以是一个包含文字的图标,其中的文字为“接机,XX公司的XXX”。虚拟对象的空间位置与第一用户的空间位置关联,并能够随着第一用户移动而移动。11 illustrates a first user and virtual objects associated with the first user as viewed by a second user through his device (eg, glasses or cell phone), according to one embodiment. The virtual object may be, for example, an icon containing text, wherein the text is "pick-up, XXX of XX company". The spatial position of the virtual object is associated with the spatial position of the first user and can move as the first user moves.
虽然在上文的一些实施例中以两个用户为例进行了说明,但这并非限制,本申请的方案同样可以适用于更多个用户。图12示出了根据一个实施例的由一个用户通过其手机屏幕观察到的实际图像,该图像中包括多个用户,每个用户具有关联的虚拟对象。Although two users are used as an example for description in some of the above embodiments, this is not a limitation, and the solution of the present application can also be applied to more users. Figure 12 shows an actual image observed by a user through his cell phone screen, the image including multiple users, each user having an associated virtual object, according to one embodiment.
在上文的实施例中,以摄像头作为传感器的示例进行描述,但可以理解,本文的实施例同样可以适用于其他任何能够感测或确定目标位置的传感器,例如激光雷达、毫米波雷达、无线信号收发器等。In the above embodiments, a camera is used as an example of a sensor for description, but it can be understood that the embodiments herein are also applicable to any other sensor that can sense or determine the target position, such as lidar, millimeter-wave radar, wireless Signal Transceivers, etc.
可以理解,本申请的实施例中涉及的设备可以是任何由用户携带或控制的设备(例如,手机、平板电脑、智能眼镜、AR眼镜、智能头盔、智能手表、车辆、等等),并且也可以是各种能够自主移动的机器,例如,无人机、无人驾驶汽车、机器人等,该设备上安装有图像采集器件。需要说明的是,本申请中的眼镜可以是AR眼镜、智能眼镜、或者任何其他能够用于向用户呈现信息的眼镜。本申请中的眼镜也包括对普通光学眼镜附加构件或插件后而形成的眼镜,例如,通过在普通光学眼镜上附加显示器件而形成的眼镜。It will be appreciated that the devices involved in the embodiments of the present application may be any devices carried or controlled by the user (eg, mobile phones, tablet computers, smart glasses, AR glasses, smart helmets, smart watches, vehicles, etc.), and also It can be various machines that can move autonomously, for example, unmanned aerial vehicles, unmanned vehicles, robots, etc., and image acquisition devices are installed on the equipment. It should be noted that the glasses in this application may be AR glasses, smart glasses, or any other glasses that can be used to present information to the user. The glasses in this application also include glasses formed by adding components or inserts to ordinary optical glasses, for example, glasses formed by adding a display device to ordinary optical glasses.
在本发明的一个实施例中,可以以计算机程序的形式来实现本发明。计算机程序可以存储于各种存储介质(例如,硬盘、光盘、闪存等)中,当该计算机程序被处理器执行时,能够用于实现本发明的方法。In one embodiment of the present invention, the present invention may be implemented in the form of a computer program. The computer program can be stored in various storage media (eg, hard disk, optical disk, flash memory, etc.), and when the computer program is executed by the processor, can be used to implement the method of the present invention.
在本发明的另一个实施例中,可以以电子设备的形式来实现本发明。该电子设备包括处理器和存储器,在存储器中存储有计算机程序,当该计算机程序被处理器执行时,能够用于实现本发明的方法。In another embodiment of the present invention, the present invention may be implemented in the form of an electronic device. The electronic device includes a processor and a memory, and the memory stores a computer program that, when executed by the processor, can be used to implement the method of the present invention.
本文中针对“各个实施例”、“一些实施例”、“一个实施例”、或“实施例”等的参考指代的是结合所述实施例所描述的特定特征、结构、或性 质包括在至少一个实施例中。因此,短语“在各个实施例中”、“在一些实施例中”、“在一个实施例中”、或“在实施例中”等在整个本文中各处的出现并非必须指代相同的实施例。此外,特定特征、结构、或性质可以在一个或多个实施例中以任何合适方式组合。因此,结合一个实施例中所示出或描述的特定特征、结构或性质可以整体地或部分地与一个或多个其他实施例的特征、结构、或性质无限制地组合,只要该组合不是不符合逻辑的或不能工作。本文中出现的类似于“根据A”、“基于A”、“通过A”或“使用A”的表述意指非排他性的,也即,“根据A”可以涵盖“仅仅根据A”,也可以涵盖“根据A和B”,除非特别声明其含义为“仅仅根据A”。在本申请中为了清楚说明,以一定的顺序描述了一些示意性的操作步骤,但本领域技术人员可以理解,这些操作步骤中的每一个并非是必不可少的,其中的一些步骤可以被省略或者被其他步骤替代。这些操作步骤也并非必须以所示的方式依次执行,相反,这些操作步骤中的一些可以根据实际需要以不同的顺序执行,或者并行执行,只要新的执行方式不是不符合逻辑的或不能工作。References herein to "various embodiments," "some embodiments," "one embodiment," or "an embodiment" etc. refer to the fact that a particular feature, structure, or property described in connection with the embodiment is included in the in at least one embodiment. Thus, appearances of the phrases "in various embodiments," "in some embodiments," "in one embodiment," or "in an embodiment" in various places throughout this document are not necessarily referring to the same implementation example. Furthermore, the particular features, structures, or properties may be combined in any suitable manner in one or more embodiments. Thus, particular features, structures, or properties shown or described in connection with one embodiment may be combined, in whole or in part, with the features, structures, or properties of one or more other embodiments without limitation, so long as the combination does not limit the Logical or not working. Expressions such as "according to A", "based on A", "by A" or "using A" appearing herein are meant to be non-exclusive, that is, "according to A" may encompass "according to A only" or Covers "according to A and B" unless specifically stated to mean "according to A only". In this application, for the sake of clarity, some schematic operation steps are described in a certain order, but those skilled in the art can understand that each of these operation steps is not essential, and some of them may be omitted or replaced by other steps. These operational steps also do not have to be performed sequentially in the manner shown, rather, some of these operational steps may be performed in a different order as practically desired, or in parallel, as long as the new implementation is not illogical or inoperable.
由此描述了本发明的至少一个实施例的几个方面,可以理解,对本领域技术人员来说容易地进行各种改变、修改和改进。这种改变、修改和改进意于在本发明的精神和范围内。虽然本发明已经通过一些实施例进行了描述,然而本发明并非局限于这里所描述的实施例,在不脱离本发明范围的情况下还包括所作出的各种改变以及变化。Having thus described several aspects of at least one embodiment of this invention, it will be appreciated that various changes, modifications, and improvements will readily occur to those skilled in the art. Such changes, modifications and improvements are intended to be within the spirit and scope of the present invention. Although the present invention has been described by way of some embodiments, the present invention is not limited to the embodiments described herein, and various changes and changes can be made without departing from the scope of the present invention.

Claims (15)

  1. 一种用于获得场景中的设备或其用户的标识信息的方法,所述场景中部署有一个或多个传感器和一个或多个视觉标志,所述传感器能够用于感测或确定所述场景中的设备或用户的位置信息,所述方法包括:A method for obtaining identification information of a device or its user in a scene in which one or more sensors and one or more visual markers are deployed, the sensors capable of being used to sense or determine the scene The location information of the device or user in the , the method includes:
    接收由设备发送的信息,该信息中包括所述设备或其用户的标识信息以及所述设备的空间位置信息,其中,所述设备通过扫描所述视觉标志来确定其空间位置信息;receiving information sent by a device, the information including identification information of the device or its user and spatial location information of the device, wherein the device determines its spatial location information by scanning the visual marker;
    基于所述设备的空间位置信息识别出在所述传感器的感测范围内的所述设备或其用户;以及Identifying the device or its user within the sensing range of the sensor based on the spatial location information of the device; and
    将所述设备或其用户的标识信息关联到在所述传感器的感测范围内的所述设备或其用户,以便向所述设备或其用户提供服务。The identification information of the device or its user is associated with the device or its user within the sensing range of the sensor, so as to provide a service to the device or its user.
  2. 根据权利要求1所述的方法,还包括:The method of claim 1, further comprising:
    通过所述传感器跟踪所述设备或其用户并更新所述设备或其用户的空间位置信息;以及Track the device or its user through the sensor and update the spatial location information of the device or its user; and
    基于所述设备或其用户的空间位置信息,向所述设备或其用户提供服务。A service is provided to the device or its user based on the spatial location information of the device or its user.
  3. 根据权利要求1所述的方法,还包括:The method of claim 1, further comprising:
    通过所述传感器跟踪所述设备或其用户并更新所述设备或其用户的空间位置信息;以及Track the device or its user through the sensor and update the spatial location information of the device or its user; and
    设置与所述设备或其用户关联的虚拟对象的相关信息,所述相关信息包括内容信息和空间位置信息,其中,所述虚拟对象的空间位置信息与所述设备或其用户的空间位置信息相关。Setting related information of the virtual object associated with the device or its user, the related information including content information and spatial location information, wherein the spatial location information of the virtual object is related to the spatial location information of the device or its user .
  4. 根据权利要3所述的方法,还包括:The method according to claim 3, further comprising:
    将所述虚拟对象的相关信息发送给其他设备,其中,所述虚拟对象的相关信息能够被所述其他设备使用以基于其位置信息和/或姿态信息在其显示媒介上呈现所述虚拟对象。Information about the virtual object is sent to the other device, wherein the information about the virtual object can be used by the other device to present the virtual object on its display medium based on its position information and/or gesture information.
  5. 根据权利要求1所述的方法,其中所述传感器包括下列一项或多 项:The method of claim 1, wherein the sensor comprises one or more of the following:
    摄像头;Camera;
    雷达;radar;
    无线信号收发器。Wireless signal transceiver.
  6. 根据权利要求1所述的方法,还包括:基于所述设备或其用户的位置信息和/或姿态信息向所述设备或其用户提供服务。The method of claim 1, further comprising: providing a service to the device or its user based on location information and/or gesture information of the device or its user.
  7. 根据权利要求6所述的方法,还包括:向所述设备发送与虚拟对象有关的信息,所述信息包括所述虚拟对象的空间位置信息,其中,所述虚拟对象能够被呈现于所述设备的显示媒介上。6. The method of claim 6, further comprising sending information about a virtual object to the device, the information including spatial location information of the virtual object, wherein the virtual object can be presented to the device on the display medium.
  8. 根据权利要求1所述的方法,还包括:The method of claim 1, further comprising:
    通过所述传感器跟踪所述设备或其用户以获得所述设备或其用户的位置信息和/或姿态信息;或者Track the device or its user through the sensor to obtain location and/or gesture information of the device or its user; or
    通过所述设备获得其位置信息和/或姿态信息。Its position information and/or attitude information is obtained through the device.
  9. 根据权利要求1所述的方法,其中,所述传感器包括摄像头,以及其中,所述基于所述设备的空间位置信息识别出在所述传感器的感测范围内的所述设备或其用户包括:The method of claim 1, wherein the sensor comprises a camera, and wherein the identifying the device or its user within the sensing range of the sensor based on the spatial location information of the device comprises:
    基于所述设备的空间位置信息确定所述设备或其用户在所述摄像头所拍摄图像中的成像位置;以及determining the imaging position of the device or its user in the image captured by the camera based on the spatial location information of the device; and
    根据所述成像位置识别出在所述摄像头所拍摄图像中的所述设备或其用户。The device or its user in the image captured by the camera is identified based on the imaging position.
  10. 根据权利要求9所述的方法,其中,所述基于所述设备的空间位置信息确定所述设备或其用户在所述摄像头所拍摄图像中的成像位置包括:The method according to claim 9, wherein the determining the imaging position of the device or its user in the image captured by the camera based on the spatial position information of the device comprises:
    基于预先建立的所述场景中的一个或多个空间位置与所述摄像头所拍摄图像中的一个或多个成像位置之间的映射关系以及所述设备的空间位置信息,来确定所述设备或其用户在所述摄像头所拍摄图像中的成像位置;或者Determine the device or the device based on the pre-established mapping relationship between one or more spatial positions in the scene and one or more imaging positions in the image captured by the camera and the spatial position information of the device. the imaging position of its user in the image captured by the camera; or
    基于所述设备的空间位置信息以及所述摄像头的位姿信息来确定所述设备或其用户在所述摄像头所拍摄图像中的成像位置。The imaging position of the device or its user in the image captured by the camera is determined based on the spatial position information of the device and the pose information of the camera.
  11. 根据权利要求1所述的方法,其中,所述基于所述设备的空间位置信息识别出在所述传感器的感测范围内的所述设备或其用户包括:The method of claim 1, wherein the identifying the device or its user within the sensing range of the sensor based on the spatial location information of the device comprises:
    将所述设备的空间位置信息与根据所述传感器的感测结果确定的一个或多个设备或用户的空间位置信息进行比较,以识别出在所述传感器的感测范围内的所述设备或其用户。comparing the spatial location information of the device with the spatial location information of one or more devices or users determined according to the sensing result of the sensor to identify the device or the user within the sensing range of the sensor its users.
  12. 根据权利要求1所述的方法,其中,所述设备通过扫描所述视觉标志来确定其空间位置信息包括:The method of claim 1, wherein the device determining its spatial location information by scanning the visual marker comprises:
    使用所述设备采集所述视觉标志的图像;capturing an image of the visual sign using the device;
    通过分析所述图像来确定所述视觉标志的标识信息以及所述设备相对于所述视觉标志的位置;determining the identification information of the visual sign and the position of the device relative to the visual sign by analyzing the image;
    通过所述视觉标志的标识信息获得所述视觉标志在空间中的位置和姿态信息;Obtain the position and attitude information of the visual sign in space through the identification information of the visual sign;
    基于所述视觉标志在空间中的位置和姿态信息以及所述设备相对于所述视觉标志的位置,确定所述设备的空间位置信息。Based on the position and attitude information of the visual landmark in space and the position of the device relative to the visual landmark, the spatial position information of the device is determined.
  13. 一种用于获得场景中的设备或其用户的标识信息的系统,所述系统包括:A system for obtaining identification information of a device in a scene or a user thereof, the system comprising:
    部署于所述场景中的一个或多个传感器,所述传感器能够用于感测或确定所述场景中的设备或用户的位置信息;one or more sensors deployed in the scene, the sensors can be used to sense or determine location information of devices or users in the scene;
    部署于所述场景中的一个或多个视觉标志;以及one or more visual markers deployed in the scene; and
    服务器,其被配置用于实现权利要求1-12中任一项所述的方法。A server configured to implement the method of any of claims 1-12.
  14. 根据权利要求13所述的系统,其中所述传感器包括下列一项或多项:The system of claim 13, wherein the sensor comprises one or more of the following:
    摄像头;Camera;
    雷达;radar;
    无线信号收发器。Wireless signal transceiver.
  15. 一种存储介质,其中存储有计算机程序,在所述计算机程序被处理器执行时,能够用于实现权利要求1-12中任一项所述的方法。A storage medium in which a computer program is stored, which can be used to implement the method of any one of claims 1-12 when the computer program is executed by a processor.
PCT/CN2021/129727 2020-12-08 2021-11-10 Method and system for obtaining identification information of device or user thereof in scenario WO2022121606A1 (en)

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
CN202011442020.6 2020-12-08
CN202011440875.5 2020-12-08
CN202011440905.2A CN112528699B (en) 2020-12-08 2020-12-08 Method and system for obtaining identification information of devices or users thereof in a scene
CN202011440875.5A CN112581630A (en) 2020-12-08 2020-12-08 User interaction method and system
CN202011442020.6A CN114663491A (en) 2020-12-08 2020-12-08 Method and system for providing information to a user in a scene
CN202011440905.2 2020-12-08

Publications (1)

Publication Number Publication Date
WO2022121606A1 true WO2022121606A1 (en) 2022-06-16

Family

ID=81973104

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/129727 WO2022121606A1 (en) 2020-12-08 2021-11-10 Method and system for obtaining identification information of device or user thereof in scenario

Country Status (2)

Country Link
TW (1) TWI800113B (en)
WO (1) WO2022121606A1 (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012182685A (en) * 2011-03-01 2012-09-20 Wham Net Service Corp Mountain entering and leaving notification system
CN103646565A (en) * 2013-12-24 2014-03-19 苏州众天力信息科技有限公司 WeChat based vehicle searching two-dimensional code position information storage and search method
CN111256701A (en) * 2020-04-26 2020-06-09 北京外号信息技术有限公司 Equipment positioning method and system
CN111814752A (en) * 2020-08-14 2020-10-23 上海木木聚枞机器人科技有限公司 Indoor positioning implementation method, server, intelligent mobile device and storage medium
CN112528699A (en) * 2020-12-08 2021-03-19 北京外号信息技术有限公司 Method and system for obtaining identification information of a device or its user in a scene
CN112581630A (en) * 2020-12-08 2021-03-30 北京外号信息技术有限公司 User interaction method and system

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9336508B2 (en) * 2012-11-02 2016-05-10 Patrick Soon-Shiong Virtual planogram management, systems, and methods

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012182685A (en) * 2011-03-01 2012-09-20 Wham Net Service Corp Mountain entering and leaving notification system
CN103646565A (en) * 2013-12-24 2014-03-19 苏州众天力信息科技有限公司 WeChat based vehicle searching two-dimensional code position information storage and search method
CN111256701A (en) * 2020-04-26 2020-06-09 北京外号信息技术有限公司 Equipment positioning method and system
CN111814752A (en) * 2020-08-14 2020-10-23 上海木木聚枞机器人科技有限公司 Indoor positioning implementation method, server, intelligent mobile device and storage medium
CN112528699A (en) * 2020-12-08 2021-03-19 北京外号信息技术有限公司 Method and system for obtaining identification information of a device or its user in a scene
CN112581630A (en) * 2020-12-08 2021-03-30 北京外号信息技术有限公司 User interaction method and system

Also Published As

Publication number Publication date
TWI800113B (en) 2023-04-21
TW202223749A (en) 2022-06-16

Similar Documents

Publication Publication Date Title
US20210019854A1 (en) Location Signaling with Respect to an Autonomous Vehicle and a Rider
KR102366293B1 (en) System and method for monitoring field based augmented reality using digital twin
CN107782314B (en) Code scanning-based augmented reality technology indoor positioning navigation method
US20180196417A1 (en) Location Signaling with Respect to an Autonomous Vehicle and a Rider
US10354407B2 (en) Camera for locating hidden objects
CN105408938B (en) System for the processing of 2D/3D space characteristics
CN105409212B (en) The electronic equipment with depth sense is caught with multi-view image
KR102289745B1 (en) System and method for real-time monitoring field work
EP3848674B1 (en) Location signaling with respect to an autonomous vehicle and a rider
EP2974509B1 (en) Personal information communicator
JP6896688B2 (en) Position calculation device, position calculation program, position calculation method, and content addition system
US10868977B2 (en) Information processing apparatus, information processing method, and program capable of adaptively displaying a video corresponding to sensed three-dimensional information
TWI750822B (en) Method and system for setting presentable virtual object for target
CN112528699B (en) Method and system for obtaining identification information of devices or users thereof in a scene
WO2022121606A1 (en) Method and system for obtaining identification information of device or user thereof in scenario
CN112788443A (en) Interaction method and system based on optical communication device
CN112581630A (en) User interaction method and system
WO2021057886A1 (en) Navigation method and system based on optical communication apparatus, and device, and medium
CN114663491A (en) Method and system for providing information to a user in a scene
TWI759764B (en) Superimpose virtual object method based on optical communitation device, electric apparatus, and computer readable storage medium
CN114071003B (en) Shooting method and system based on optical communication device
US20220084258A1 (en) Interaction method based on optical communication apparatus, and electronic device
CN114827338A (en) Method and electronic device for presenting virtual objects on a display medium of a device
CN112561953A (en) Method and system for target recognition and tracking in real scenes
CN111752293A (en) Method and electronic device for guiding a machine capable of autonomous movement

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21902313

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21902313

Country of ref document: EP

Kind code of ref document: A1