WO2015187335A1 - Device-provided tracking data for augmented reality - Google Patents

Device-provided tracking data for augmented reality Download PDF

Info

Publication number
WO2015187335A1
WO2015187335A1 PCT/US2015/030859 US2015030859W WO2015187335A1 WO 2015187335 A1 WO2015187335 A1 WO 2015187335A1 US 2015030859 W US2015030859 W US 2015030859W WO 2015187335 A1 WO2015187335 A1 WO 2015187335A1
Authority
WO
WIPO (PCT)
Prior art keywords
tracking data
networked device
networked
augmented reality
query
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/US2015/030859
Other languages
English (en)
French (fr)
Inventor
Gabriel Berelejis
Ravinder Paul Chandhok
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Qualcomm Inc
Original Assignee
Qualcomm Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Qualcomm Inc filed Critical Qualcomm Inc
Priority to JP2016570282A priority Critical patent/JP6416290B2/ja
Priority to CN201580028761.0A priority patent/CN106462322B/zh
Priority to EP15728249.2A priority patent/EP3149564A1/en
Publication of WO2015187335A1 publication Critical patent/WO2015187335A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/04815Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/20Scenes; Scene-specific elements in augmented reality scenes
    • GPHYSICS
    • G08SIGNALLING
    • G08CTRANSMISSION SYSTEMS FOR MEASURED VALUES, CONTROL OR SIMILAR SIGNALS
    • G08C17/00Arrangements for transmitting signals characterised by the use of a wireless electrical link
    • G08C17/02Arrangements for transmitting signals characterised by the use of a wireless electrical link using a radio link
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/12Protocols specially adapted for proprietary or special-purpose networking environments, e.g. medical networks, sensor networks, networks in vehicles or remote metering networks
    • H04L67/125Protocols specially adapted for proprietary or special-purpose networking environments, e.g. medical networks, sensor networks, networks in vehicles or remote metering networks involving control of end-device applications over a network
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/535Tracking the activity of the user
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q9/00Arrangements in telecontrol or telemetry systems for selectively calling a substation from a main station, in which substation desired apparatus is selected for applying a control signal thereto or for obtaining measured values therefrom
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72409User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories
    • H04M1/72412User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories using two-way short-range wireless interfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72427User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for supporting games or graphical animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q2209/00Arrangements in telecontrol or telemetry systems
    • H04Q2209/40Arrangements in telecontrol or telemetry systems using a wireless architecture
    • H04Q2209/43Arrangements in telecontrol or telemetry systems using a wireless architecture using wireless personal area networks [WPAN], e.g. 802.15, 802.15.1, 802.15.4, Bluetooth or ZigBee

Definitions

  • Embodiments of the disclosed subject matter generally relate to the field of electronic devices, and, more particularly, to augmented reality.
  • Augmented reality is a technology that enhances a live view of the real world with virtual, computer-generated objects such as stationary and/or moving images, text, data, etc.
  • AR Augmented reality
  • a camera view of a physical, real-world environment may be augmented (or
  • An AR device refers to a device which runs an AR application.
  • An AR application may augment a camera view to show information about objects in the camera view. For example, some AR applications overlay information about points of interest and geographic landmarks. More recently, AR applications can augment a camera view to provide information or user interface actions associated with nearby objects.
  • an AR application may recognize an object (sometimes referred to as a trackable object) in the camera view and augment the camera view with additional information or user interface actions related to the object. To recognize objects in the camera view, the AR application may require tracking data which describes the object to be tracked. Once an object is identified in a camera view using the tracking data, the AR application can augment the camera view with additional information about the object.
  • an AR device may recognize a trackable object which can communicate via a local network.
  • An object that can communicate via a local network may be referred to as a networked device.
  • Tracking data that describes the networked device may be provided by the networked device via the local network to the AR device.
  • the tracking data may be embedded in the networked device.
  • a method performed by an augmented reality (AR) device comprises receiving tracking data that describes one or more trackable features of a networked device, wherein the tracking data is received from the networked device via a local network; and generating an augmented reality view in association with the networked device in response to detecting at least one trackable feature of the networked device in a camera view of the AR device.
  • AR augmented reality
  • the method further comprises determining, based at least in part on the tracking data, a position of the networked device in the camera view of the AR device.
  • generating the augmented reality view includes augmenting at least a portion of the augmented reality view with AR properties associated with the networked device.
  • the tracking data describes one or more visual properties of at least a first trackable feature, the one or more visual properties including of at least one of size, shape, and color.
  • receiving the tracking data comprises sending a query to the networked device to request the tracking data from the networked device; and receiving the tracking data in response to the query.
  • the method further comprises determining that the networked device is in proximity of the AR device prior to sending the query to the networked device.
  • receiving the tracking data comprises receiving a broadcast message from the networked device, the broadcast message including the tracking data.
  • the local network comprises a wireless local area network (WLAN).
  • WLAN wireless local area network
  • the tracking data is received via a device interoperability framework message.
  • the tracking data is received via a short-range radio frequency communication medium.
  • the tracking data is included in an augmented reality property data set associated with the networked device, the augmented reality property data set embedded in the networked device.
  • the method further comprises obtaining the camera view using an integrated camera at the AR device.
  • the method further comprises displaying the augmented reality view using an integrated display apparatus at the AR device.
  • an augmented reality (AR) device comprises a processor; and memory storing instructions, which when executed by the processor, cause the AR device to receive tracking data that describes one or more trackable features of a networked device, wherein the tracking data is received from the networked device via a local network; and generate an augmented reality view in association with the networked device in response to detecting at least one trackable feature of the networked device in a camera view of the AR device.
  • AR augmented reality
  • the instructions further cause the processor to determine, based at least in part on the tracking data, a position of the networked device in the camera view of the AR device.
  • the instructions further cause the processor to execute an AR application that generates the augmented reality view based on the camera view, wherein the augmented reality view is augmented with AR properties associated with the networked device.
  • the AR device further comprises a display apparatus that displays the augmented reality view to a user of the AR device.
  • the AR device further comprises a network interface that sends a query to the networked device to request the tracking data from the networked device, and that receives the tracking data in response to the query.
  • the network interface sends the query in response to the AR device determining that the networked device is in proximity of the AR device.
  • the AR device further comprises a network interface that receives a broadcast message from the networked device, the broadcast message including the tracking data.
  • a method performed by a networked device comprises storing tracking data that describes one or more trackable features of the networked device, the tracking data enabling visual detection of the networked device by an augmented reality (AR) device; and providing the tracking data to the AR device via a local network.
  • AR augmented reality
  • the tracking data describes one or more visual properties of at least a first trackable feature, the one or more visual properties including of at least one of size, shape, and color.
  • the method further comprises receiving a query from the AR device requesting the tracking data, wherein providing the tracking data is in response to the query.
  • the method further comprises providing an indication to the AR device that the networked device has the tracking data stored at the networked device.
  • the method further comprises broadcasting the tracking data via the local network in accordance with a device discovery protocol.
  • the tracking data is included in an augmented reality data set stored at the networked device and provided to the AR device in response to a request from the AR device.
  • the tracking data is provided via a device interoperability framework message.
  • a networked device comprises a network interface to communicate via a local network; a processor; and a memory storing tracking data that describes one or more trackable features of the networked device, the tracking data enabling visual detection of the networked device by an augmented reality (AR) device, the memory storing instructions, which when executed by the processor, cause the networked device to provide the tracking data to the AR device via the local network.
  • AR augmented reality
  • the instructions cause the processor to receive a query from the AR device requesting the tracking data, and provide the tracking data is in response to the query.
  • the instructions cause the processor to provide an indication to the AR device that the networked device has the tracking data stored at the networked device.
  • Figure 1 depicts an example system to introduce various concepts of this disclosure.
  • Figure 2 depicts example operations using tracking data for augmented reality.
  • Figure 3 depicts example components that may be used in various embodiments of this disclosure.
  • Figure 4 depicts a message flow of example messages between a networked device and an AR device in a local network in accordance with an embodiment of this disclosure.
  • Figure 5 depicts example operations performed by a networked device to provide tracking data in accordance with an embodiment of this disclosure.
  • Figure 6 depicts an example electronic device capable of implementing various embodiments of this disclosure.
  • tracking data may also describe other properties a trackable feature, including auditory, infrared, patterns, or other properties which may be sensed by the AR device.
  • well-known instruction instances, protocols, structures and techniques have not been shown in detail in order not to obfuscate the description.
  • An AR device may include a processor, memory, and a display housed in an electronic device.
  • the AR device can be any suitable electronic device, e.g., desktop, laptop, hand-held, mobile, or wearable computer.
  • a camera may be integrated in the AR device or may be external to the AR device. The camera may provide a live view, sometimes referred to as the "camera view," to the AR device.
  • the AR device may execute an AR application to determine the position of the camera view in the real world and detect any objects (e.g., AR-enabled objects) in the camera view. When an object is detected, the AR application may generate an augmented reality view based on the camera view and computer-generated overlay information (such as virtual objects, images, user interface controls, etc.) related to the object.
  • the overlay information may be referred to as AR properties.
  • the AR device may use tracking data that describes an object to be tracked.
  • Tracking data may describe a variety of trackable features (sometimes referred to as "trackables") based on patterns that can be searched, recognized and/or tracked in the camera view.
  • tracking data may describe fixed markers or may describe characteristics for markerless detection. Fixed markers are known characteristics that can be identified using pattern matching or image matching.
  • the tracking data may include interest points, fiduciary markers, quick response (QR) codes, images to be matched, or any "trackable" pattern.
  • Markers can be of any design, now known or developed in the future, including a circular, linear, matrix, variable bit length matrix, multi-level matrix, black/white (binary), gray scale patterns, and combinations thereof.
  • the markers can be two- or three-dimensional barcodes.
  • an AR device can detect an object based on descriptive characteristics rather than a fixed marker.
  • the tracking data may include descriptions based on edges, points, corners, flat surfaces, colors, sizes, shapes, geometry, etc.
  • the tracking data may include a multi-dimensional (e.g., 2D or 3D) model of an object so that the AR device can detect the object regardless of which angle the camera view has in relation to the object.
  • the tracking data may be hard coded in an AR application or may be downloaded from a cloud server via a wide area network.
  • a new object may be introduced into an environment, and the AR application may not have tracking data associated with the new object. It may be undesirable to download tracking data from a network-based server during an augmented reality session or for each new object introduced into an environment.
  • a local network may include a wireless local area network (WLAN) (such as IEEE 802.11a/b/g/n/ac/ad), short-range radio frequency medium (such as BluetoothTM), near field communication (NFC), powerline communication (PLC), Ethernet, or the like.
  • WLAN wireless local area network
  • BluetoothTM short-range radio frequency medium
  • NFC near field communication
  • PLC powerline communication
  • Ethernet or the like.
  • an AR device and a networked device may communicate using a protocol to support interoperability (which may be referred to as a device interoperability framework).
  • IoT Internet of Things
  • various devices may communicate using a communication protocol, such as AllJoynTM.
  • Other protocols may be used by the various devices.
  • a device interoperability framework protocol can be any protocol that allows an AR device and a networked device to communicate via a local network.
  • a networked device may provide tracking data associated with the networked device to an AR device via the local network.
  • the networked device can store tracking data and can provide the tracking data directly to the AR device via the local network.
  • the tracking data may describe trackable features that enable the AR device to detect the networked device in a camera view.
  • an AR device may obtain the tracking data from the networked device using the device interoperability framework protocol. Because the tracking data may be obtained via the local network, the AR device may be the tracking data without communicating with a wide area network server or without having the tracking data previously coded into an AR application.
  • the tracking data may be in the form of a local protocol extension, AR markup language, or object description property.
  • FIG. 1 depicts an example system 100 to introduce various concepts of this disclosure.
  • the example system 100 includes a networked device 1 10, a local network 130, and an AR device 120.
  • a networked device 1 10 is depicted as a smart appliance - such as a stove.
  • the networked device 110 may be any type of device which is AR- enabled and that can communicate via the local network 130.
  • the local network 130 may be a powerline communication network, a wireless local area network, short range radio frequency network, or the like.
  • the AR device 120 can also communicate via the local network 130.
  • the AR device 120 is depicted as a tablet computer.
  • the AR device 120 and the networked device 1 10 may be of different forms or machines, and the examples in Figure 1 are provided for illustrative purposes.
  • the networked device 1 10 is configured to provide tracking data (shown as arrow 145) to the AR device 120 via the local network 130.
  • the networked device 1 10 may communicate the tracking data using a device interoperability framework message.
  • message may be an AllJoynTM protocol message or another suitable interoperability framework message associated with a message syntax.
  • the local network 130 comprises a wireless local area network providing a wireless coverage range. Devices such as the networked device 110 and AR device 120 that are associated with the wireless local area network may communicate using the device interoperability framework protocol.
  • the networked device 110 may store the tracking data in local memory or in a storage device of the networked device 1 10 and provide the tracking data to the AR device 120 in a device interoperability framework protocol message. In another embodiment, the networked device 110 may provide the tracking data using a standardized or extensible protocol message.
  • the AR device 120 may receive the tracking data from the networked device 110 and use the tracking data to detect the networked device 1 10 in a camera view of the AR device 120.
  • the AR device 120 may analyze the camera view to detect edges, feature points, colors, shapes, sizes, etc., as may be described in the tracking data.
  • the AR device 120 may detect (shown as arrow 155) the networked device 110 based on the tracking data.
  • the tracking data may describe the color, size, shape, or geometry associated with the control knobs of the stove.
  • control panel may be a trackable feature of the networked device 1 10.
  • the knobs may represent feature points, and the distance between the feature points may be described by the tracking data.
  • the AR device 120 may detect a pattern consistent with the knobs of the stove as described in the tracking data.
  • the AR device 120 may register the networked device 110 as a trackable object in an augmented reality view.
  • the networked device 1 10 is in the camera view 165, and the AR device 120 has enhanced the camera view to include an AR property 175 associated with the networked device 110.
  • the AR device 120 may generate an augmented reality view that includes the images of the real world enhanced with overlay information associated with the detected AR objects (i.e., networked device 110).
  • the overlay information may be predefined or obtained via a wide area network server.
  • the overlay information may also be provided by the networked device 110, similar to the tracking data 145.
  • the AR device 120 includes a display apparatus 122 that displays the augmented reality view to a user of the AR device 120.
  • one or both of the camera and the display apparatus 122 may be integrated into the AR device 120. However, in other embodiments, one or both of the camera and the display apparatus 122 may be external to the AR device 120. For example, one or both of the camera and display apparatus may be part of a wearable accessory, such as electronic glasses worn by a user. The camera and/or display apparatus may be external components of a computer system that comprises the AR device 120. In another embodiment, the AR device 120 may be part of a home network that includes a camera and/or display (e.g., television) that are in communication with the AR device 120.
  • a camera and/or display e.g., television
  • FIG. 2 depicts example operations (flowchart 200) using tracking data for augmented reality.
  • the AR device may receive tracking data that describes one or more trackable features of a networked device, wherein the tracking data is received from the networked device via a local network.
  • the AR device may determine, based at least in part on the tracking data, a position of the networked device in a camera view at the AR device.
  • the AR device may generate an augmented reality view in association with the networked device in response to detecting at least one trackable feature of the networked device in a camera view of the AR device.
  • the AR device may display the augmented reality view.
  • the augmented reality view may be displayed using an integrated display apparatus at the AR device or an external display in communication with the AR device.
  • FIG. 3 depicts example components that may be used in one embodiment of this disclosure.
  • an example system 300 includes a networked device 110 and an AR device 120, both coupled to a local network 130.
  • the networked device 110 may include a communication unit 316, a controller 314, and tracking data 312.
  • the controller 314 may control other aspects (not shown) of the networked device 1 10.
  • the controller 314 may control a temperature setting of the stove in Figure 1.
  • Other examples in which a controller 314 may control aspects of the networked device 110 may be conceived. Examples may include control of power, lighting, temperature, intensity, etc.
  • the tracking data 312 may be stored in a memory (not shown) associated with the networked device 1 10.
  • the communication unit 316 may comprise a network interface to couple the networked device 110 to the local network 130.
  • the communication unit 316 may implement a device interoperability framework protocol which can be used to convey the tracking data 312 to the AR device 120.
  • the controller 314 may be remotely controllable using an AR application 323 of the AR device 120.
  • the AR device 120 of Figure 3 includes a camera 322, AR application 323, display 324, and a communication unit 326.
  • the communication unit 326 may comprise a network interface and may couple the AR device 120 to the local network 130.
  • the AR application 323 may be configured to receive the tracking data from the networked device 110.
  • the AR application 323 may process the tracking data as needed to enable detection of the networked device 110 by the AR application 323.
  • the AR application 323 may use a camera view from the camera 322 to generate an augmented reality view.
  • the augmented reality view may be output by the display 324.
  • the camera or other video input may generate a digitized video image of the real world, which is referred to as a camera view in this disclosure.
  • the camera may be any digital device now known or developed in the future that is dimensioned and configured to capture still or motion pictures of the real world and to convert those images to a digital stream of information that can be manipulated by the AR device 120.
  • the camera may be a digital still camera, a digital video camera, a web cam, a head-mounted display, a camera phone, a tablet personal computer, or any camera that renders a digital image or stream of images over which augmented reality data may be added.
  • the tracking data may describe one or more trackable features associated with the networked device.
  • the trackable feature may also be an image of a real world item which the AR application 323 can recognize.
  • the AR application 323 can recognize a refrigerator or other appliance in a video stream of a kitchen based on an image of the refrigerator in the tracking data.
  • the tracking data may be used to describe features of the networked device 1 10 when no markers are present.
  • the tracking data may describe the color, shape, size, and geometry, etc. regarding one or more trackable features of the networked device 110.
  • FIG. 4 depicts a message flow 400 of example messages between a networked device 110 and an AR device 120 using a device interoperability framework protocol.
  • the AR device 120 may become aware that the networked device 1 10 has tracking data to provide to the AR device 120. There may be a number of ways that the AR device 120 determines that the networked device 1 10 can provide the tracking data.
  • the networked device 110 may transmit a broadcast message 410 that includes an indicator that tracking data is available.
  • the broadcast message 410 may include some or all of the tracking data as part of the broadcast message 410.
  • the broadcast message 410 may be part of a device discovery protocol.
  • the AR device 120 may send an AR enablement query message 420 to the networked device 1 10 to inquire whether the networked device 110 is AR-enabled.
  • the AR enablement query message 420 may request, for example, the
  • AR TrackingDataAvailable property, or the like, from the networked device 1 10.
  • Other augmented reality properties may be solicited in the AR enablement query message 420.
  • the networked device 1 10 may respond to the AR enablement query message 420 by sending an AR enablement response message 430 to indicate whether the networked device 110 is AR-enabled.
  • the AR enablement response message 430 may include a property of
  • AR_TrackingDataAvailable Yes" or similar property.
  • the AR device 120 may attempt to obtain tracking data associated with the networked device 1 10.
  • the AR device 120 may transmit an AR tracking data query message 440 to the networked device 110.
  • the networked device 1 10 may respond with an AR tracking data response message 450 that includes the tracking data.
  • the AR device 120 may use the tracking data to detect and register the networked device 110 in a camera view of the AR device 120. Once detected and registered, the AR device 120 may display overlay information (such as AR properties, controls, etc.) associated with the networked device 110 when the networked device 110 is positioned in the camera view of the AR device 120. For example, the AR device 120 may generate an augmented reality view based on the camera view and overlay information associated with the networked device 110. The augmented reality view can then be displayed by the AR device 120.
  • overlay information such as AR properties, controls, etc.
  • FIG. 5 depicts example operations (flowchart 500) performed by a networked device to provide tracking data in accordance with an embodiment of this disclosure.
  • the networked device may store tracking data that describes one or more trackable features of the networked device, the tracking data enabling visual detection of the networked device by an augmented reality (AR) device.
  • AR augmented reality
  • the memory or storage component may be integrated with the networked device, or may be an external memory storage communicatively coupled with the networked device.
  • the networked device may optionally inform (e.g., via an indication in a protocol message or a broadcast message) the AR device that the networked device has the tracking data stored at the networked device.
  • the networked device may receive a query from the AR device requesting the tracking data. In some embodiments, the networked device may not receive a query from the AR device, but may instead simply broadcast the tracking data periodically.
  • the networked device may provide the tracking data to the AR device via a local network in response to the query.
  • the tracking data may be provided using a structured markup language in accordance with an embodiment of this disclosure. It will be appreciated that a variety of formats can be used to describe the tracking data, such as Augmented Reality Markup Language (ARML).
  • a structured markup language file may be modified to include tracking data.
  • a section of the structured markup language file may describe an object (e.g., networked device) that should be tracked.
  • Another section of the structured markup language file may define one or more trackable features of the object that should be tracked.
  • a trackable feature may be a front control panel of the stove.
  • the structured markup language file may include a tracking data field which contains the tracking data associated with the trackable feature.
  • the tracking data may describe size, shape, color, edges, feature points, or other aspects of the trackable feature.
  • the format of the tracking data may be text, binary data or any suitable format, and may be compressed or encrypted in some embodiments.
  • the tracking data may also include modeling data such as a 2D or 3D model of the trackable feature.
  • the structured markup language may include other AR properties (not shown) which further enhance the capabilities of the AR device to overlay information in relation with the networked device.
  • aspects of the present disclosure may be embodied as a system, method, or computer program product. Accordingly, aspects of the present disclosure may take the form of an entirely hardware embodiment, a software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a "circuit,” “module” or “system.” Furthermore, aspects of the present disclosure may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
  • Non-transitory computer-readable media comprise all computer-readable media, with the sole exception being a transitory, propagating signal.
  • the non-transitory computer readable medium may be a computer readable storage medium.
  • a computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing.
  • a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • Computer program code for carrying out operations of the embodiments may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages.
  • the program code may execute entirely on a user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN), a personal area network (PAN), or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • LAN local area network
  • PAN personal area network
  • WAN wide area network
  • Internet Service Provider an Internet Service Provider
  • FIG. 6 is an example block diagram of one embodiment of an electronic device 600 capable of implementing various embodiments of this disclosure.
  • the electronic device 600 may be an electronic device such as a laptop computer, a tablet computer, a mobile phone, a powerline communication device, a gaming console, or other electronic systems.
  • the electronic device may comprise functionality to communicate across multiple communication networks (which form a hybrid communication network).
  • the electronic device 600 includes a processor 602 (possibly including multiple processors, multiple cores, multiple nodes, and/or implementing multi-threading, etc.).
  • the electronic device 600 includes a memory 606.
  • the memory 606 may be system memory (e.g., one or more of cache, SRAM, DRAM, zero capacitor RAM, Twin Transistor RAM, eDRAM, EDO RAM, DDR RAM, EEPROM, NRAM, RRAM, SONOS, PRAM, etc.) or any one or more of the above already described possible realizations of machine-readable media.
  • the electronic device 600 also includes a bus 610 (e.g., PCI, ISA, PCI-Express, HyperTransport®,
  • the one or more network interfaces 604 may include a wireless network interface (e.g., a WLAN interface, a Bluetooth® interface, a WiMAX interface, a ZigBee® interface, a Wireless USB interface, etc.) or a wired network interface (e.g., a powerline communication interface, an Ethernet interface, etc.).
  • a wireless network interface e.g., a WLAN interface, a Bluetooth® interface, a WiMAX interface, a ZigBee® interface, a Wireless USB interface, etc.
  • a wired network interface e.g., a powerline communication interface, an Ethernet interface, etc.
  • the electronic device 600 includes an AR application 623.
  • the electronic device 600 may include one or both of a camera 622 and a display apparatus 624.
  • AR application 623 is depicted as a separate hardware component of the electronic device 600, it should be noted that the AR application 623 may be implemented as machine readable instructions stored in the memory 606 and executed by the processor 602.
  • the AR application 623 may implement various embodiments of the foregoing figures.
  • the AR application 623, camera 622, and display apparatus 624 may operate similar to the corresponding components described in Figure 3.
  • the electronic device 600 may also be used to describe the networked device (such as networked device 1 10). However, the networked device may not have a camera 622, display apparatus 624, or AR application 623 in some embodiments.
  • the electronic device 600 may store the tracking data in the memory 606 and provide the tracking data via the local network using network interface 604.
  • any one of these functionalities may be partially (or entirely) implemented in hardware and/or on the processor 602.
  • the functionality may be implemented with an application specific integrated circuit, in logic implemented in the processor 602, in a coprocessor on a peripheral device or card, etc.
  • realizations may include fewer or additional components not illustrated in Figure 11 (e.g., video cards, audio cards, additional network interfaces, peripheral devices, etc.).
  • the processor 602, the memory 606, network interface(s) 604 may be coupled to the bus 610. Although illustrated as being coupled to the bus 610, the memory 606 may be directly coupled to the processor 602.

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Computer Hardware Design (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Computer Graphics (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)
  • Information Transfer Between Computers (AREA)
PCT/US2015/030859 2014-06-02 2015-05-14 Device-provided tracking data for augmented reality Ceased WO2015187335A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
JP2016570282A JP6416290B2 (ja) 2014-06-02 2015-05-14 拡張現実のためにデバイスが提供する追跡データ
CN201580028761.0A CN106462322B (zh) 2014-06-02 2015-05-14 用于增强现实的设备提供的跟踪数据
EP15728249.2A EP3149564A1 (en) 2014-06-02 2015-05-14 Device-provided tracking data for augmented reality

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201462006749P 2014-06-02 2014-06-02
US62/006,749 2014-06-02
US14/507,591 US9886162B2 (en) 2014-06-02 2014-10-06 Device-provided tracking data for augmented reality
US14/507,591 2014-10-06

Publications (1)

Publication Number Publication Date
WO2015187335A1 true WO2015187335A1 (en) 2015-12-10

Family

ID=54702164

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2015/030859 Ceased WO2015187335A1 (en) 2014-06-02 2015-05-14 Device-provided tracking data for augmented reality

Country Status (5)

Country Link
US (1) US9886162B2 (enExample)
EP (1) EP3149564A1 (enExample)
JP (1) JP6416290B2 (enExample)
CN (1) CN106462322B (enExample)
WO (1) WO2015187335A1 (enExample)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2017120329A (ja) * 2015-12-28 2017-07-06 株式会社ブリリアントサービス 調理用ヘッドマウントディスプレイおよび調理用ヘッドマウントディスプレイのプログラム
US10499997B2 (en) 2017-01-03 2019-12-10 Mako Surgical Corp. Systems and methods for surgical navigation
US11280801B2 (en) 2019-01-28 2022-03-22 Becton, Dickinson And Company Hazardous contaminant collection device with integrated swab and test device
US11360001B2 (en) 2017-09-21 2022-06-14 Becton, Dickinson And Company Reactive demarcation template for hazardous contaminant testing
US11380074B2 (en) 2017-09-21 2022-07-05 Becton, Dickinson And Company Augmented reality devices for hazardous contaminant testing
US11385146B2 (en) 2017-09-21 2022-07-12 Becton, Dickinson And Company Sampling systems and techniques to collect hazardous contaminants with high pickup and shedding efficiencies
US11391748B2 (en) 2017-09-21 2022-07-19 Becton, Dickinson And Company High dynamic range assays in hazardous contaminant testing
US11585733B2 (en) 2017-09-21 2023-02-21 Becton, Dickinson And Company Hazardous contaminant collection kit and rapid testing
US11782042B2 (en) 2017-09-21 2023-10-10 Becton, Dickinson And Company Hazardous contaminant collection kit and rapid testing
US11821819B2 (en) 2017-09-21 2023-11-21 Becton, Dickinson And Company Demarcation template for hazardous contaminant testing

Families Citing this family (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016077506A1 (en) 2014-11-11 2016-05-19 Bent Image Lab, Llc Accurate positioning of augmented reality content
US10600249B2 (en) 2015-10-16 2020-03-24 Youar Inc. Augmented reality platform
KR102479578B1 (ko) * 2016-02-03 2022-12-20 삼성전자주식회사 전자장치 및 그 제어방법
US10802695B2 (en) * 2016-03-23 2020-10-13 Youar Inc. Augmented reality for the internet of things
US10110678B2 (en) * 2016-08-19 2018-10-23 Sony Corporation System and method for data communication based on image processing
US11782570B2 (en) * 2017-02-23 2023-10-10 Jue-Hsuan Hsiao Integration platform of internet of things and virtual device
WO2018164532A1 (en) * 2017-03-09 2018-09-13 Samsung Electronics Co., Ltd. System and method for enhancing augmented reality (ar) experience on user equipment (ue) based on in-device contents
CN107067428B (zh) * 2017-03-10 2020-06-30 深圳奥比中光科技有限公司 增强现实投影装置及方法
CN106998283A (zh) * 2017-03-28 2017-08-01 朱正和 一种基于光纤传输的远程控制vr系统
US10602046B2 (en) 2017-07-11 2020-03-24 Htc Corporation Mobile device and control method
CN109931923B (zh) 2017-12-15 2023-07-07 阿里巴巴集团控股有限公司 一种导航引导图的生成方法和装置
US10679414B2 (en) * 2017-12-15 2020-06-09 Oath Inc. Presenting an augmented reality interface
US10410426B2 (en) * 2017-12-19 2019-09-10 GM Global Technology Operations LLC Augmented reality vehicle user interface
US10726463B2 (en) * 2017-12-20 2020-07-28 Signify Holding B.V. Lighting and internet of things design using augmented reality
US10225360B1 (en) 2018-01-24 2019-03-05 Veeva Systems Inc. System and method for distributing AR content
CN108551420B (zh) * 2018-04-08 2021-12-14 北京灵犀微光科技有限公司 增强现实设备及其信息处理方法
CN110568824A (zh) * 2018-06-05 2019-12-13 鑀錹科技有限公司 工业物联网架构
US11163434B2 (en) * 2019-01-24 2021-11-02 Ademco Inc. Systems and methods for using augmenting reality to control a connected home system
US11206505B2 (en) 2019-05-06 2021-12-21 Universal City Studios Llc Systems and methods for dynamically loading area-based augmented reality content
WO2020239832A1 (en) * 2019-05-31 2020-12-03 BSH Hausgeräte GmbH Method and device for providing information regarding a gas hob
IT201900015485A1 (it) * 2019-09-03 2021-03-03 Ali Group Srl Carpigiani Sistema di supporto per la gestione di una macchina per il trattamento di prodotti alimentari e procedimento corrispondente.
US11533434B2 (en) * 2019-09-18 2022-12-20 Google Llc Generating and rendering motion graphics effects based on recognized content in camera view finder
CN113970888A (zh) * 2020-07-07 2022-01-25 华为技术有限公司 家居设备控制方法、终端设备及计算机可读存储介质
US11385071B2 (en) 2020-08-07 2022-07-12 Micron Technology, Inc. Providing a route with augmented reality
US11954885B2 (en) * 2021-09-15 2024-04-09 Apple Inc. Display tracking systems and methods
US11809680B2 (en) * 2021-12-30 2023-11-07 Snap Inc. Interface for engaging IoT devices with AR camera
MX2024008032A (es) * 2021-12-30 2024-07-15 Assurant Inc Aparatos, metodos implementados por computadora, y productos de programas de computadora para el procesamiento de realidad aumentada centrado en la red.
US11855831B1 (en) 2022-06-10 2023-12-26 T-Mobile Usa, Inc. Enabling an operator to resolve an issue associated with a 5G wireless telecommunication network using AR glasses
US11886767B2 (en) 2022-06-17 2024-01-30 T-Mobile Usa, Inc. Enable interaction between a user and an agent of a 5G wireless telecommunication network using augmented reality glasses

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2293531A1 (en) * 2009-08-11 2011-03-09 Lg Electronics Inc. Electronic device and control method thereof
WO2013154476A1 (en) * 2012-04-12 2013-10-17 Telefonaktiebolaget L M Ericsson (Publ) Pairing a mobile terminal with a wireless device

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5731785A (en) * 1994-05-13 1998-03-24 Lemelson; Jerome H. System and method for locating objects including an inhibiting feature
US7047525B2 (en) * 2001-04-02 2006-05-16 American Express Travel Related Services Company, Inc. System and method for an interoperability framework
US7127082B2 (en) 2002-09-27 2006-10-24 Hrl Laboratories, Llc Active fiducials for augmented reality
US8180396B2 (en) 2007-10-18 2012-05-15 Yahoo! Inc. User augmented reality for camera-enabled mobile devices
US8606657B2 (en) 2009-01-21 2013-12-10 Edgenet, Inc. Augmented reality method and system for designing environments and buying/selling goods
KR101633359B1 (ko) 2009-10-20 2016-06-27 삼성전자 주식회사 투사 불변량을 이용한 무표식 증강 현실 구현 시스템 및 그 방법
US8442502B2 (en) 2010-03-02 2013-05-14 Empire Technology Development, Llc Tracking an object in augmented reality
JP4914528B1 (ja) * 2010-08-31 2012-04-11 新日鉄ソリューションズ株式会社 拡張現実提供システム、情報処理端末、情報処理装置、拡張現実提供方法、情報処理方法、及びプログラム
JP2012152016A (ja) * 2011-01-19 2012-08-09 Mitsubishi Electric Corp ケーブル接続作業支援装置
US9480913B2 (en) * 2011-01-26 2016-11-01 WhitewaterWest Industries Ltd. Interactive entertainment using a mobile device with object tagging and/or hyperlinking
US20130155107A1 (en) 2011-12-16 2013-06-20 Identive Group, Inc. Systems and Methods for Providing an Augmented Reality Experience
US9350814B2 (en) 2012-02-21 2016-05-24 Qualcomm Incorporated Internet protocol connectivity over a service-oriented architecture bus
US9143402B2 (en) 2012-02-24 2015-09-22 Qualcomm Incorporated Sensor based configuration and control of network devices
US9449343B2 (en) 2012-10-05 2016-09-20 Sap Se Augmented-reality shopping using a networked mobile device
US9081994B2 (en) * 2012-10-05 2015-07-14 Hand Held Products, Inc. Portable RFID reading terminal with visual indication of scan trace
US9740935B2 (en) * 2013-11-26 2017-08-22 Honeywell International Inc. Maintenance assistant system

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2293531A1 (en) * 2009-08-11 2011-03-09 Lg Electronics Inc. Electronic device and control method thereof
WO2013154476A1 (en) * 2012-04-12 2013-10-17 Telefonaktiebolaget L M Ericsson (Publ) Pairing a mobile terminal with a wireless device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3149564A1 *

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2017120329A (ja) * 2015-12-28 2017-07-06 株式会社ブリリアントサービス 調理用ヘッドマウントディスプレイおよび調理用ヘッドマウントディスプレイのプログラム
US10499997B2 (en) 2017-01-03 2019-12-10 Mako Surgical Corp. Systems and methods for surgical navigation
US12383347B2 (en) 2017-01-03 2025-08-12 Mako Surgical Corp. Systems and methods for surgical navigation
US11707330B2 (en) 2017-01-03 2023-07-25 Mako Surgical Corp. Systems and methods for surgical navigation
US11391748B2 (en) 2017-09-21 2022-07-19 Becton, Dickinson And Company High dynamic range assays in hazardous contaminant testing
US11385146B2 (en) 2017-09-21 2022-07-12 Becton, Dickinson And Company Sampling systems and techniques to collect hazardous contaminants with high pickup and shedding efficiencies
US11380074B2 (en) 2017-09-21 2022-07-05 Becton, Dickinson And Company Augmented reality devices for hazardous contaminant testing
US11585733B2 (en) 2017-09-21 2023-02-21 Becton, Dickinson And Company Hazardous contaminant collection kit and rapid testing
US11360001B2 (en) 2017-09-21 2022-06-14 Becton, Dickinson And Company Reactive demarcation template for hazardous contaminant testing
US11782042B2 (en) 2017-09-21 2023-10-10 Becton, Dickinson And Company Hazardous contaminant collection kit and rapid testing
US11821819B2 (en) 2017-09-21 2023-11-21 Becton, Dickinson And Company Demarcation template for hazardous contaminant testing
US12399088B2 (en) 2017-09-21 2025-08-26 Becton, Dickinson And Company Demarcation template for hazardous contaminant testing
US12436158B2 (en) 2017-09-21 2025-10-07 Becton, Dickinson And Company High dynamic range assays in hazardous contaminant testing
US11860173B2 (en) 2019-01-28 2024-01-02 Becton, Dickinson And Company Hazardous contaminant collection device with integrated swab and test device
US11280801B2 (en) 2019-01-28 2022-03-22 Becton, Dickinson And Company Hazardous contaminant collection device with integrated swab and test device

Also Published As

Publication number Publication date
US20150347850A1 (en) 2015-12-03
EP3149564A1 (en) 2017-04-05
US9886162B2 (en) 2018-02-06
JP2017527870A (ja) 2017-09-21
JP6416290B2 (ja) 2018-10-31
CN106462322B (zh) 2019-08-06
CN106462322A (zh) 2017-02-22

Similar Documents

Publication Publication Date Title
US9886162B2 (en) Device-provided tracking data for augmented reality
US8872852B2 (en) Positional context determination with multi marker confidence ranking
CN108307214B (zh) 用于控制装置的方法和设备
US9430877B2 (en) Electronic device and method for selecting augmented content using the same
KR102418992B1 (ko) 전자 장치 및 전자 장치의 증강 현실 서비스 제공 방법
KR102479360B1 (ko) 증강 현실 서비스 제공 방법 및 장치
CN103365414A (zh) 信息处理设备、信息处理方法和信息处理系统
KR20120133648A (ko) 증강현실 컨텐츠를 통해 타겟장치를 제어하는 제어단말기 및 서버
CN104982090A (zh) 个人信息通信器
US20170171521A1 (en) Projection apparatus and operation method thereof
CN106131794A (zh) 室内定位的方法及装置
CN108540542A (zh) 一种移动增强现实系统及显示的方法
EP3229482B1 (en) Master device, slave device, and control method therefor
CN117716253A (zh) 映射联网设备
US10652041B2 (en) Computer vision based activation
WO2017088331A1 (zh) 控制装置及运作方法
KR101840381B1 (ko) 다중 디바이스 연동 제어 시스템
CN107330974B (zh) 商品展示方法、装置及移动设备
JP2012123548A5 (enExample)
CN108615260A (zh) 一种异常现实环境下进行增强现实数字文化内容显示的方法及装置
AU2018341597B2 (en) Using augmented reality for secure transactions
US20200009791A1 (en) Smart handling of materials for three-dimensional (3d) printers
Park et al. A Study on Intuitive IoT Interface System using 3D Depth Camera
US20220417479A1 (en) Information processing device and information processing method
JP7111416B2 (ja) 携帯端末、情報処理システム、制御方法、及びプログラム

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15728249

Country of ref document: EP

Kind code of ref document: A1

DPE1 Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101)
REEP Request for entry into the european phase

Ref document number: 2015728249

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2015728249

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 2016570282

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE