WO2015187335A1 - Device-provided tracking data for augmented reality - Google Patents
Device-provided tracking data for augmented reality Download PDFInfo
- Publication number
- WO2015187335A1 WO2015187335A1 PCT/US2015/030859 US2015030859W WO2015187335A1 WO 2015187335 A1 WO2015187335 A1 WO 2015187335A1 US 2015030859 W US2015030859 W US 2015030859W WO 2015187335 A1 WO2015187335 A1 WO 2015187335A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- tracking data
- networked device
- networked
- augmented reality
- query
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04815—Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/20—Scenes; Scene-specific elements in augmented reality scenes
-
- G—PHYSICS
- G08—SIGNALLING
- G08C—TRANSMISSION SYSTEMS FOR MEASURED VALUES, CONTROL OR SIMILAR SIGNALS
- G08C17/00—Arrangements for transmitting signals characterised by the use of a wireless electrical link
- G08C17/02—Arrangements for transmitting signals characterised by the use of a wireless electrical link using a radio link
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/12—Protocols specially adapted for proprietary or special-purpose networking environments, e.g. medical networks, sensor networks, networks in vehicles or remote metering networks
- H04L67/125—Protocols specially adapted for proprietary or special-purpose networking environments, e.g. medical networks, sensor networks, networks in vehicles or remote metering networks involving control of end-device applications over a network
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/50—Network services
- H04L67/535—Tracking the activity of the user
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04Q—SELECTING
- H04Q9/00—Arrangements in telecontrol or telemetry systems for selectively calling a substation from a main station, in which substation desired apparatus is selected for applying a control signal thereto or for obtaining measured values therefrom
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72403—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
- H04M1/72409—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories
- H04M1/72412—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories using two-way short-range wireless interfaces
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72403—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
- H04M1/72427—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for supporting games or graphical animations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04Q—SELECTING
- H04Q2209/00—Arrangements in telecontrol or telemetry systems
- H04Q2209/40—Arrangements in telecontrol or telemetry systems using a wireless architecture
- H04Q2209/43—Arrangements in telecontrol or telemetry systems using a wireless architecture using wireless personal area networks [WPAN], e.g. 802.15, 802.15.1, 802.15.4, Bluetooth or ZigBee
Definitions
- Embodiments of the disclosed subject matter generally relate to the field of electronic devices, and, more particularly, to augmented reality.
- Augmented reality is a technology that enhances a live view of the real world with virtual, computer-generated objects such as stationary and/or moving images, text, data, etc.
- AR Augmented reality
- a camera view of a physical, real-world environment may be augmented (or
- An AR device refers to a device which runs an AR application.
- An AR application may augment a camera view to show information about objects in the camera view. For example, some AR applications overlay information about points of interest and geographic landmarks. More recently, AR applications can augment a camera view to provide information or user interface actions associated with nearby objects.
- an AR application may recognize an object (sometimes referred to as a trackable object) in the camera view and augment the camera view with additional information or user interface actions related to the object. To recognize objects in the camera view, the AR application may require tracking data which describes the object to be tracked. Once an object is identified in a camera view using the tracking data, the AR application can augment the camera view with additional information about the object.
- an AR device may recognize a trackable object which can communicate via a local network.
- An object that can communicate via a local network may be referred to as a networked device.
- Tracking data that describes the networked device may be provided by the networked device via the local network to the AR device.
- the tracking data may be embedded in the networked device.
- a method performed by an augmented reality (AR) device comprises receiving tracking data that describes one or more trackable features of a networked device, wherein the tracking data is received from the networked device via a local network; and generating an augmented reality view in association with the networked device in response to detecting at least one trackable feature of the networked device in a camera view of the AR device.
- AR augmented reality
- the method further comprises determining, based at least in part on the tracking data, a position of the networked device in the camera view of the AR device.
- generating the augmented reality view includes augmenting at least a portion of the augmented reality view with AR properties associated with the networked device.
- the tracking data describes one or more visual properties of at least a first trackable feature, the one or more visual properties including of at least one of size, shape, and color.
- receiving the tracking data comprises sending a query to the networked device to request the tracking data from the networked device; and receiving the tracking data in response to the query.
- the method further comprises determining that the networked device is in proximity of the AR device prior to sending the query to the networked device.
- receiving the tracking data comprises receiving a broadcast message from the networked device, the broadcast message including the tracking data.
- the local network comprises a wireless local area network (WLAN).
- WLAN wireless local area network
- the tracking data is received via a device interoperability framework message.
- the tracking data is received via a short-range radio frequency communication medium.
- the tracking data is included in an augmented reality property data set associated with the networked device, the augmented reality property data set embedded in the networked device.
- the method further comprises obtaining the camera view using an integrated camera at the AR device.
- the method further comprises displaying the augmented reality view using an integrated display apparatus at the AR device.
- an augmented reality (AR) device comprises a processor; and memory storing instructions, which when executed by the processor, cause the AR device to receive tracking data that describes one or more trackable features of a networked device, wherein the tracking data is received from the networked device via a local network; and generate an augmented reality view in association with the networked device in response to detecting at least one trackable feature of the networked device in a camera view of the AR device.
- AR augmented reality
- the instructions further cause the processor to determine, based at least in part on the tracking data, a position of the networked device in the camera view of the AR device.
- the instructions further cause the processor to execute an AR application that generates the augmented reality view based on the camera view, wherein the augmented reality view is augmented with AR properties associated with the networked device.
- the AR device further comprises a display apparatus that displays the augmented reality view to a user of the AR device.
- the AR device further comprises a network interface that sends a query to the networked device to request the tracking data from the networked device, and that receives the tracking data in response to the query.
- the network interface sends the query in response to the AR device determining that the networked device is in proximity of the AR device.
- the AR device further comprises a network interface that receives a broadcast message from the networked device, the broadcast message including the tracking data.
- a method performed by a networked device comprises storing tracking data that describes one or more trackable features of the networked device, the tracking data enabling visual detection of the networked device by an augmented reality (AR) device; and providing the tracking data to the AR device via a local network.
- AR augmented reality
- the tracking data describes one or more visual properties of at least a first trackable feature, the one or more visual properties including of at least one of size, shape, and color.
- the method further comprises receiving a query from the AR device requesting the tracking data, wherein providing the tracking data is in response to the query.
- the method further comprises providing an indication to the AR device that the networked device has the tracking data stored at the networked device.
- the method further comprises broadcasting the tracking data via the local network in accordance with a device discovery protocol.
- the tracking data is included in an augmented reality data set stored at the networked device and provided to the AR device in response to a request from the AR device.
- the tracking data is provided via a device interoperability framework message.
- a networked device comprises a network interface to communicate via a local network; a processor; and a memory storing tracking data that describes one or more trackable features of the networked device, the tracking data enabling visual detection of the networked device by an augmented reality (AR) device, the memory storing instructions, which when executed by the processor, cause the networked device to provide the tracking data to the AR device via the local network.
- AR augmented reality
- the instructions cause the processor to receive a query from the AR device requesting the tracking data, and provide the tracking data is in response to the query.
- the instructions cause the processor to provide an indication to the AR device that the networked device has the tracking data stored at the networked device.
- Figure 1 depicts an example system to introduce various concepts of this disclosure.
- Figure 2 depicts example operations using tracking data for augmented reality.
- Figure 3 depicts example components that may be used in various embodiments of this disclosure.
- Figure 4 depicts a message flow of example messages between a networked device and an AR device in a local network in accordance with an embodiment of this disclosure.
- Figure 5 depicts example operations performed by a networked device to provide tracking data in accordance with an embodiment of this disclosure.
- Figure 6 depicts an example electronic device capable of implementing various embodiments of this disclosure.
- tracking data may also describe other properties a trackable feature, including auditory, infrared, patterns, or other properties which may be sensed by the AR device.
- well-known instruction instances, protocols, structures and techniques have not been shown in detail in order not to obfuscate the description.
- An AR device may include a processor, memory, and a display housed in an electronic device.
- the AR device can be any suitable electronic device, e.g., desktop, laptop, hand-held, mobile, or wearable computer.
- a camera may be integrated in the AR device or may be external to the AR device. The camera may provide a live view, sometimes referred to as the "camera view," to the AR device.
- the AR device may execute an AR application to determine the position of the camera view in the real world and detect any objects (e.g., AR-enabled objects) in the camera view. When an object is detected, the AR application may generate an augmented reality view based on the camera view and computer-generated overlay information (such as virtual objects, images, user interface controls, etc.) related to the object.
- the overlay information may be referred to as AR properties.
- the AR device may use tracking data that describes an object to be tracked.
- Tracking data may describe a variety of trackable features (sometimes referred to as "trackables") based on patterns that can be searched, recognized and/or tracked in the camera view.
- tracking data may describe fixed markers or may describe characteristics for markerless detection. Fixed markers are known characteristics that can be identified using pattern matching or image matching.
- the tracking data may include interest points, fiduciary markers, quick response (QR) codes, images to be matched, or any "trackable" pattern.
- Markers can be of any design, now known or developed in the future, including a circular, linear, matrix, variable bit length matrix, multi-level matrix, black/white (binary), gray scale patterns, and combinations thereof.
- the markers can be two- or three-dimensional barcodes.
- an AR device can detect an object based on descriptive characteristics rather than a fixed marker.
- the tracking data may include descriptions based on edges, points, corners, flat surfaces, colors, sizes, shapes, geometry, etc.
- the tracking data may include a multi-dimensional (e.g., 2D or 3D) model of an object so that the AR device can detect the object regardless of which angle the camera view has in relation to the object.
- the tracking data may be hard coded in an AR application or may be downloaded from a cloud server via a wide area network.
- a new object may be introduced into an environment, and the AR application may not have tracking data associated with the new object. It may be undesirable to download tracking data from a network-based server during an augmented reality session or for each new object introduced into an environment.
- a local network may include a wireless local area network (WLAN) (such as IEEE 802.11a/b/g/n/ac/ad), short-range radio frequency medium (such as BluetoothTM), near field communication (NFC), powerline communication (PLC), Ethernet, or the like.
- WLAN wireless local area network
- BluetoothTM short-range radio frequency medium
- NFC near field communication
- PLC powerline communication
- Ethernet or the like.
- an AR device and a networked device may communicate using a protocol to support interoperability (which may be referred to as a device interoperability framework).
- IoT Internet of Things
- various devices may communicate using a communication protocol, such as AllJoynTM.
- Other protocols may be used by the various devices.
- a device interoperability framework protocol can be any protocol that allows an AR device and a networked device to communicate via a local network.
- a networked device may provide tracking data associated with the networked device to an AR device via the local network.
- the networked device can store tracking data and can provide the tracking data directly to the AR device via the local network.
- the tracking data may describe trackable features that enable the AR device to detect the networked device in a camera view.
- an AR device may obtain the tracking data from the networked device using the device interoperability framework protocol. Because the tracking data may be obtained via the local network, the AR device may be the tracking data without communicating with a wide area network server or without having the tracking data previously coded into an AR application.
- the tracking data may be in the form of a local protocol extension, AR markup language, or object description property.
- FIG. 1 depicts an example system 100 to introduce various concepts of this disclosure.
- the example system 100 includes a networked device 1 10, a local network 130, and an AR device 120.
- a networked device 1 10 is depicted as a smart appliance - such as a stove.
- the networked device 110 may be any type of device which is AR- enabled and that can communicate via the local network 130.
- the local network 130 may be a powerline communication network, a wireless local area network, short range radio frequency network, or the like.
- the AR device 120 can also communicate via the local network 130.
- the AR device 120 is depicted as a tablet computer.
- the AR device 120 and the networked device 1 10 may be of different forms or machines, and the examples in Figure 1 are provided for illustrative purposes.
- the networked device 1 10 is configured to provide tracking data (shown as arrow 145) to the AR device 120 via the local network 130.
- the networked device 1 10 may communicate the tracking data using a device interoperability framework message.
- message may be an AllJoynTM protocol message or another suitable interoperability framework message associated with a message syntax.
- the local network 130 comprises a wireless local area network providing a wireless coverage range. Devices such as the networked device 110 and AR device 120 that are associated with the wireless local area network may communicate using the device interoperability framework protocol.
- the networked device 110 may store the tracking data in local memory or in a storage device of the networked device 1 10 and provide the tracking data to the AR device 120 in a device interoperability framework protocol message. In another embodiment, the networked device 110 may provide the tracking data using a standardized or extensible protocol message.
- the AR device 120 may receive the tracking data from the networked device 110 and use the tracking data to detect the networked device 1 10 in a camera view of the AR device 120.
- the AR device 120 may analyze the camera view to detect edges, feature points, colors, shapes, sizes, etc., as may be described in the tracking data.
- the AR device 120 may detect (shown as arrow 155) the networked device 110 based on the tracking data.
- the tracking data may describe the color, size, shape, or geometry associated with the control knobs of the stove.
- control panel may be a trackable feature of the networked device 1 10.
- the knobs may represent feature points, and the distance between the feature points may be described by the tracking data.
- the AR device 120 may detect a pattern consistent with the knobs of the stove as described in the tracking data.
- the AR device 120 may register the networked device 110 as a trackable object in an augmented reality view.
- the networked device 1 10 is in the camera view 165, and the AR device 120 has enhanced the camera view to include an AR property 175 associated with the networked device 110.
- the AR device 120 may generate an augmented reality view that includes the images of the real world enhanced with overlay information associated with the detected AR objects (i.e., networked device 110).
- the overlay information may be predefined or obtained via a wide area network server.
- the overlay information may also be provided by the networked device 110, similar to the tracking data 145.
- the AR device 120 includes a display apparatus 122 that displays the augmented reality view to a user of the AR device 120.
- one or both of the camera and the display apparatus 122 may be integrated into the AR device 120. However, in other embodiments, one or both of the camera and the display apparatus 122 may be external to the AR device 120. For example, one or both of the camera and display apparatus may be part of a wearable accessory, such as electronic glasses worn by a user. The camera and/or display apparatus may be external components of a computer system that comprises the AR device 120. In another embodiment, the AR device 120 may be part of a home network that includes a camera and/or display (e.g., television) that are in communication with the AR device 120.
- a camera and/or display e.g., television
- FIG. 2 depicts example operations (flowchart 200) using tracking data for augmented reality.
- the AR device may receive tracking data that describes one or more trackable features of a networked device, wherein the tracking data is received from the networked device via a local network.
- the AR device may determine, based at least in part on the tracking data, a position of the networked device in a camera view at the AR device.
- the AR device may generate an augmented reality view in association with the networked device in response to detecting at least one trackable feature of the networked device in a camera view of the AR device.
- the AR device may display the augmented reality view.
- the augmented reality view may be displayed using an integrated display apparatus at the AR device or an external display in communication with the AR device.
- FIG. 3 depicts example components that may be used in one embodiment of this disclosure.
- an example system 300 includes a networked device 110 and an AR device 120, both coupled to a local network 130.
- the networked device 110 may include a communication unit 316, a controller 314, and tracking data 312.
- the controller 314 may control other aspects (not shown) of the networked device 1 10.
- the controller 314 may control a temperature setting of the stove in Figure 1.
- Other examples in which a controller 314 may control aspects of the networked device 110 may be conceived. Examples may include control of power, lighting, temperature, intensity, etc.
- the tracking data 312 may be stored in a memory (not shown) associated with the networked device 1 10.
- the communication unit 316 may comprise a network interface to couple the networked device 110 to the local network 130.
- the communication unit 316 may implement a device interoperability framework protocol which can be used to convey the tracking data 312 to the AR device 120.
- the controller 314 may be remotely controllable using an AR application 323 of the AR device 120.
- the AR device 120 of Figure 3 includes a camera 322, AR application 323, display 324, and a communication unit 326.
- the communication unit 326 may comprise a network interface and may couple the AR device 120 to the local network 130.
- the AR application 323 may be configured to receive the tracking data from the networked device 110.
- the AR application 323 may process the tracking data as needed to enable detection of the networked device 110 by the AR application 323.
- the AR application 323 may use a camera view from the camera 322 to generate an augmented reality view.
- the augmented reality view may be output by the display 324.
- the camera or other video input may generate a digitized video image of the real world, which is referred to as a camera view in this disclosure.
- the camera may be any digital device now known or developed in the future that is dimensioned and configured to capture still or motion pictures of the real world and to convert those images to a digital stream of information that can be manipulated by the AR device 120.
- the camera may be a digital still camera, a digital video camera, a web cam, a head-mounted display, a camera phone, a tablet personal computer, or any camera that renders a digital image or stream of images over which augmented reality data may be added.
- the tracking data may describe one or more trackable features associated with the networked device.
- the trackable feature may also be an image of a real world item which the AR application 323 can recognize.
- the AR application 323 can recognize a refrigerator or other appliance in a video stream of a kitchen based on an image of the refrigerator in the tracking data.
- the tracking data may be used to describe features of the networked device 1 10 when no markers are present.
- the tracking data may describe the color, shape, size, and geometry, etc. regarding one or more trackable features of the networked device 110.
- FIG. 4 depicts a message flow 400 of example messages between a networked device 110 and an AR device 120 using a device interoperability framework protocol.
- the AR device 120 may become aware that the networked device 1 10 has tracking data to provide to the AR device 120. There may be a number of ways that the AR device 120 determines that the networked device 1 10 can provide the tracking data.
- the networked device 110 may transmit a broadcast message 410 that includes an indicator that tracking data is available.
- the broadcast message 410 may include some or all of the tracking data as part of the broadcast message 410.
- the broadcast message 410 may be part of a device discovery protocol.
- the AR device 120 may send an AR enablement query message 420 to the networked device 1 10 to inquire whether the networked device 110 is AR-enabled.
- the AR enablement query message 420 may request, for example, the
- AR TrackingDataAvailable property, or the like, from the networked device 1 10.
- Other augmented reality properties may be solicited in the AR enablement query message 420.
- the networked device 1 10 may respond to the AR enablement query message 420 by sending an AR enablement response message 430 to indicate whether the networked device 110 is AR-enabled.
- the AR enablement response message 430 may include a property of
- AR_TrackingDataAvailable Yes" or similar property.
- the AR device 120 may attempt to obtain tracking data associated with the networked device 1 10.
- the AR device 120 may transmit an AR tracking data query message 440 to the networked device 110.
- the networked device 1 10 may respond with an AR tracking data response message 450 that includes the tracking data.
- the AR device 120 may use the tracking data to detect and register the networked device 110 in a camera view of the AR device 120. Once detected and registered, the AR device 120 may display overlay information (such as AR properties, controls, etc.) associated with the networked device 110 when the networked device 110 is positioned in the camera view of the AR device 120. For example, the AR device 120 may generate an augmented reality view based on the camera view and overlay information associated with the networked device 110. The augmented reality view can then be displayed by the AR device 120.
- overlay information such as AR properties, controls, etc.
- FIG. 5 depicts example operations (flowchart 500) performed by a networked device to provide tracking data in accordance with an embodiment of this disclosure.
- the networked device may store tracking data that describes one or more trackable features of the networked device, the tracking data enabling visual detection of the networked device by an augmented reality (AR) device.
- AR augmented reality
- the memory or storage component may be integrated with the networked device, or may be an external memory storage communicatively coupled with the networked device.
- the networked device may optionally inform (e.g., via an indication in a protocol message or a broadcast message) the AR device that the networked device has the tracking data stored at the networked device.
- the networked device may receive a query from the AR device requesting the tracking data. In some embodiments, the networked device may not receive a query from the AR device, but may instead simply broadcast the tracking data periodically.
- the networked device may provide the tracking data to the AR device via a local network in response to the query.
- the tracking data may be provided using a structured markup language in accordance with an embodiment of this disclosure. It will be appreciated that a variety of formats can be used to describe the tracking data, such as Augmented Reality Markup Language (ARML).
- a structured markup language file may be modified to include tracking data.
- a section of the structured markup language file may describe an object (e.g., networked device) that should be tracked.
- Another section of the structured markup language file may define one or more trackable features of the object that should be tracked.
- a trackable feature may be a front control panel of the stove.
- the structured markup language file may include a tracking data field which contains the tracking data associated with the trackable feature.
- the tracking data may describe size, shape, color, edges, feature points, or other aspects of the trackable feature.
- the format of the tracking data may be text, binary data or any suitable format, and may be compressed or encrypted in some embodiments.
- the tracking data may also include modeling data such as a 2D or 3D model of the trackable feature.
- the structured markup language may include other AR properties (not shown) which further enhance the capabilities of the AR device to overlay information in relation with the networked device.
- aspects of the present disclosure may be embodied as a system, method, or computer program product. Accordingly, aspects of the present disclosure may take the form of an entirely hardware embodiment, a software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a "circuit,” “module” or “system.” Furthermore, aspects of the present disclosure may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
- Non-transitory computer-readable media comprise all computer-readable media, with the sole exception being a transitory, propagating signal.
- the non-transitory computer readable medium may be a computer readable storage medium.
- a computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing.
- a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
- Computer program code for carrying out operations of the embodiments may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages.
- the program code may execute entirely on a user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
- the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN), a personal area network (PAN), or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- LAN local area network
- PAN personal area network
- WAN wide area network
- Internet Service Provider an Internet Service Provider
- FIG. 6 is an example block diagram of one embodiment of an electronic device 600 capable of implementing various embodiments of this disclosure.
- the electronic device 600 may be an electronic device such as a laptop computer, a tablet computer, a mobile phone, a powerline communication device, a gaming console, or other electronic systems.
- the electronic device may comprise functionality to communicate across multiple communication networks (which form a hybrid communication network).
- the electronic device 600 includes a processor 602 (possibly including multiple processors, multiple cores, multiple nodes, and/or implementing multi-threading, etc.).
- the electronic device 600 includes a memory 606.
- the memory 606 may be system memory (e.g., one or more of cache, SRAM, DRAM, zero capacitor RAM, Twin Transistor RAM, eDRAM, EDO RAM, DDR RAM, EEPROM, NRAM, RRAM, SONOS, PRAM, etc.) or any one or more of the above already described possible realizations of machine-readable media.
- the electronic device 600 also includes a bus 610 (e.g., PCI, ISA, PCI-Express, HyperTransport®,
- the one or more network interfaces 604 may include a wireless network interface (e.g., a WLAN interface, a Bluetooth® interface, a WiMAX interface, a ZigBee® interface, a Wireless USB interface, etc.) or a wired network interface (e.g., a powerline communication interface, an Ethernet interface, etc.).
- a wireless network interface e.g., a WLAN interface, a Bluetooth® interface, a WiMAX interface, a ZigBee® interface, a Wireless USB interface, etc.
- a wired network interface e.g., a powerline communication interface, an Ethernet interface, etc.
- the electronic device 600 includes an AR application 623.
- the electronic device 600 may include one or both of a camera 622 and a display apparatus 624.
- AR application 623 is depicted as a separate hardware component of the electronic device 600, it should be noted that the AR application 623 may be implemented as machine readable instructions stored in the memory 606 and executed by the processor 602.
- the AR application 623 may implement various embodiments of the foregoing figures.
- the AR application 623, camera 622, and display apparatus 624 may operate similar to the corresponding components described in Figure 3.
- the electronic device 600 may also be used to describe the networked device (such as networked device 1 10). However, the networked device may not have a camera 622, display apparatus 624, or AR application 623 in some embodiments.
- the electronic device 600 may store the tracking data in the memory 606 and provide the tracking data via the local network using network interface 604.
- any one of these functionalities may be partially (or entirely) implemented in hardware and/or on the processor 602.
- the functionality may be implemented with an application specific integrated circuit, in logic implemented in the processor 602, in a coprocessor on a peripheral device or card, etc.
- realizations may include fewer or additional components not illustrated in Figure 11 (e.g., video cards, audio cards, additional network interfaces, peripheral devices, etc.).
- the processor 602, the memory 606, network interface(s) 604 may be coupled to the bus 610. Although illustrated as being coupled to the bus 610, the memory 606 may be directly coupled to the processor 602.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Computer Hardware Design (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Computer Graphics (AREA)
- Software Systems (AREA)
- Multimedia (AREA)
- User Interface Of Digital Computer (AREA)
- Information Transfer Between Computers (AREA)
Priority Applications (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2016570282A JP6416290B2 (ja) | 2014-06-02 | 2015-05-14 | 拡張現実のためにデバイスが提供する追跡データ |
| CN201580028761.0A CN106462322B (zh) | 2014-06-02 | 2015-05-14 | 用于增强现实的设备提供的跟踪数据 |
| EP15728249.2A EP3149564A1 (en) | 2014-06-02 | 2015-05-14 | Device-provided tracking data for augmented reality |
Applications Claiming Priority (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US201462006749P | 2014-06-02 | 2014-06-02 | |
| US62/006,749 | 2014-06-02 | ||
| US14/507,591 US9886162B2 (en) | 2014-06-02 | 2014-10-06 | Device-provided tracking data for augmented reality |
| US14/507,591 | 2014-10-06 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2015187335A1 true WO2015187335A1 (en) | 2015-12-10 |
Family
ID=54702164
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/US2015/030859 Ceased WO2015187335A1 (en) | 2014-06-02 | 2015-05-14 | Device-provided tracking data for augmented reality |
Country Status (5)
| Country | Link |
|---|---|
| US (1) | US9886162B2 (enExample) |
| EP (1) | EP3149564A1 (enExample) |
| JP (1) | JP6416290B2 (enExample) |
| CN (1) | CN106462322B (enExample) |
| WO (1) | WO2015187335A1 (enExample) |
Cited By (10)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2017120329A (ja) * | 2015-12-28 | 2017-07-06 | 株式会社ブリリアントサービス | 調理用ヘッドマウントディスプレイおよび調理用ヘッドマウントディスプレイのプログラム |
| US10499997B2 (en) | 2017-01-03 | 2019-12-10 | Mako Surgical Corp. | Systems and methods for surgical navigation |
| US11280801B2 (en) | 2019-01-28 | 2022-03-22 | Becton, Dickinson And Company | Hazardous contaminant collection device with integrated swab and test device |
| US11360001B2 (en) | 2017-09-21 | 2022-06-14 | Becton, Dickinson And Company | Reactive demarcation template for hazardous contaminant testing |
| US11380074B2 (en) | 2017-09-21 | 2022-07-05 | Becton, Dickinson And Company | Augmented reality devices for hazardous contaminant testing |
| US11385146B2 (en) | 2017-09-21 | 2022-07-12 | Becton, Dickinson And Company | Sampling systems and techniques to collect hazardous contaminants with high pickup and shedding efficiencies |
| US11391748B2 (en) | 2017-09-21 | 2022-07-19 | Becton, Dickinson And Company | High dynamic range assays in hazardous contaminant testing |
| US11585733B2 (en) | 2017-09-21 | 2023-02-21 | Becton, Dickinson And Company | Hazardous contaminant collection kit and rapid testing |
| US11782042B2 (en) | 2017-09-21 | 2023-10-10 | Becton, Dickinson And Company | Hazardous contaminant collection kit and rapid testing |
| US11821819B2 (en) | 2017-09-21 | 2023-11-21 | Becton, Dickinson And Company | Demarcation template for hazardous contaminant testing |
Families Citing this family (29)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2016077506A1 (en) | 2014-11-11 | 2016-05-19 | Bent Image Lab, Llc | Accurate positioning of augmented reality content |
| US10600249B2 (en) | 2015-10-16 | 2020-03-24 | Youar Inc. | Augmented reality platform |
| KR102479578B1 (ko) * | 2016-02-03 | 2022-12-20 | 삼성전자주식회사 | 전자장치 및 그 제어방법 |
| US10802695B2 (en) * | 2016-03-23 | 2020-10-13 | Youar Inc. | Augmented reality for the internet of things |
| US10110678B2 (en) * | 2016-08-19 | 2018-10-23 | Sony Corporation | System and method for data communication based on image processing |
| US11782570B2 (en) * | 2017-02-23 | 2023-10-10 | Jue-Hsuan Hsiao | Integration platform of internet of things and virtual device |
| WO2018164532A1 (en) * | 2017-03-09 | 2018-09-13 | Samsung Electronics Co., Ltd. | System and method for enhancing augmented reality (ar) experience on user equipment (ue) based on in-device contents |
| CN107067428B (zh) * | 2017-03-10 | 2020-06-30 | 深圳奥比中光科技有限公司 | 增强现实投影装置及方法 |
| CN106998283A (zh) * | 2017-03-28 | 2017-08-01 | 朱正和 | 一种基于光纤传输的远程控制vr系统 |
| US10602046B2 (en) | 2017-07-11 | 2020-03-24 | Htc Corporation | Mobile device and control method |
| CN109931923B (zh) | 2017-12-15 | 2023-07-07 | 阿里巴巴集团控股有限公司 | 一种导航引导图的生成方法和装置 |
| US10679414B2 (en) * | 2017-12-15 | 2020-06-09 | Oath Inc. | Presenting an augmented reality interface |
| US10410426B2 (en) * | 2017-12-19 | 2019-09-10 | GM Global Technology Operations LLC | Augmented reality vehicle user interface |
| US10726463B2 (en) * | 2017-12-20 | 2020-07-28 | Signify Holding B.V. | Lighting and internet of things design using augmented reality |
| US10225360B1 (en) | 2018-01-24 | 2019-03-05 | Veeva Systems Inc. | System and method for distributing AR content |
| CN108551420B (zh) * | 2018-04-08 | 2021-12-14 | 北京灵犀微光科技有限公司 | 增强现实设备及其信息处理方法 |
| CN110568824A (zh) * | 2018-06-05 | 2019-12-13 | 鑀錹科技有限公司 | 工业物联网架构 |
| US11163434B2 (en) * | 2019-01-24 | 2021-11-02 | Ademco Inc. | Systems and methods for using augmenting reality to control a connected home system |
| US11206505B2 (en) | 2019-05-06 | 2021-12-21 | Universal City Studios Llc | Systems and methods for dynamically loading area-based augmented reality content |
| WO2020239832A1 (en) * | 2019-05-31 | 2020-12-03 | BSH Hausgeräte GmbH | Method and device for providing information regarding a gas hob |
| IT201900015485A1 (it) * | 2019-09-03 | 2021-03-03 | Ali Group Srl Carpigiani | Sistema di supporto per la gestione di una macchina per il trattamento di prodotti alimentari e procedimento corrispondente. |
| US11533434B2 (en) * | 2019-09-18 | 2022-12-20 | Google Llc | Generating and rendering motion graphics effects based on recognized content in camera view finder |
| CN113970888A (zh) * | 2020-07-07 | 2022-01-25 | 华为技术有限公司 | 家居设备控制方法、终端设备及计算机可读存储介质 |
| US11385071B2 (en) | 2020-08-07 | 2022-07-12 | Micron Technology, Inc. | Providing a route with augmented reality |
| US11954885B2 (en) * | 2021-09-15 | 2024-04-09 | Apple Inc. | Display tracking systems and methods |
| US11809680B2 (en) * | 2021-12-30 | 2023-11-07 | Snap Inc. | Interface for engaging IoT devices with AR camera |
| MX2024008032A (es) * | 2021-12-30 | 2024-07-15 | Assurant Inc | Aparatos, metodos implementados por computadora, y productos de programas de computadora para el procesamiento de realidad aumentada centrado en la red. |
| US11855831B1 (en) | 2022-06-10 | 2023-12-26 | T-Mobile Usa, Inc. | Enabling an operator to resolve an issue associated with a 5G wireless telecommunication network using AR glasses |
| US11886767B2 (en) | 2022-06-17 | 2024-01-30 | T-Mobile Usa, Inc. | Enable interaction between a user and an agent of a 5G wireless telecommunication network using augmented reality glasses |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| EP2293531A1 (en) * | 2009-08-11 | 2011-03-09 | Lg Electronics Inc. | Electronic device and control method thereof |
| WO2013154476A1 (en) * | 2012-04-12 | 2013-10-17 | Telefonaktiebolaget L M Ericsson (Publ) | Pairing a mobile terminal with a wireless device |
Family Cites Families (16)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US5731785A (en) * | 1994-05-13 | 1998-03-24 | Lemelson; Jerome H. | System and method for locating objects including an inhibiting feature |
| US7047525B2 (en) * | 2001-04-02 | 2006-05-16 | American Express Travel Related Services Company, Inc. | System and method for an interoperability framework |
| US7127082B2 (en) | 2002-09-27 | 2006-10-24 | Hrl Laboratories, Llc | Active fiducials for augmented reality |
| US8180396B2 (en) | 2007-10-18 | 2012-05-15 | Yahoo! Inc. | User augmented reality for camera-enabled mobile devices |
| US8606657B2 (en) | 2009-01-21 | 2013-12-10 | Edgenet, Inc. | Augmented reality method and system for designing environments and buying/selling goods |
| KR101633359B1 (ko) | 2009-10-20 | 2016-06-27 | 삼성전자 주식회사 | 투사 불변량을 이용한 무표식 증강 현실 구현 시스템 및 그 방법 |
| US8442502B2 (en) | 2010-03-02 | 2013-05-14 | Empire Technology Development, Llc | Tracking an object in augmented reality |
| JP4914528B1 (ja) * | 2010-08-31 | 2012-04-11 | 新日鉄ソリューションズ株式会社 | 拡張現実提供システム、情報処理端末、情報処理装置、拡張現実提供方法、情報処理方法、及びプログラム |
| JP2012152016A (ja) * | 2011-01-19 | 2012-08-09 | Mitsubishi Electric Corp | ケーブル接続作業支援装置 |
| US9480913B2 (en) * | 2011-01-26 | 2016-11-01 | WhitewaterWest Industries Ltd. | Interactive entertainment using a mobile device with object tagging and/or hyperlinking |
| US20130155107A1 (en) | 2011-12-16 | 2013-06-20 | Identive Group, Inc. | Systems and Methods for Providing an Augmented Reality Experience |
| US9350814B2 (en) | 2012-02-21 | 2016-05-24 | Qualcomm Incorporated | Internet protocol connectivity over a service-oriented architecture bus |
| US9143402B2 (en) | 2012-02-24 | 2015-09-22 | Qualcomm Incorporated | Sensor based configuration and control of network devices |
| US9449343B2 (en) | 2012-10-05 | 2016-09-20 | Sap Se | Augmented-reality shopping using a networked mobile device |
| US9081994B2 (en) * | 2012-10-05 | 2015-07-14 | Hand Held Products, Inc. | Portable RFID reading terminal with visual indication of scan trace |
| US9740935B2 (en) * | 2013-11-26 | 2017-08-22 | Honeywell International Inc. | Maintenance assistant system |
-
2014
- 2014-10-06 US US14/507,591 patent/US9886162B2/en active Active
-
2015
- 2015-05-14 WO PCT/US2015/030859 patent/WO2015187335A1/en not_active Ceased
- 2015-05-14 CN CN201580028761.0A patent/CN106462322B/zh active Active
- 2015-05-14 EP EP15728249.2A patent/EP3149564A1/en not_active Ceased
- 2015-05-14 JP JP2016570282A patent/JP6416290B2/ja not_active Expired - Fee Related
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| EP2293531A1 (en) * | 2009-08-11 | 2011-03-09 | Lg Electronics Inc. | Electronic device and control method thereof |
| WO2013154476A1 (en) * | 2012-04-12 | 2013-10-17 | Telefonaktiebolaget L M Ericsson (Publ) | Pairing a mobile terminal with a wireless device |
Non-Patent Citations (1)
| Title |
|---|
| See also references of EP3149564A1 * |
Cited By (15)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2017120329A (ja) * | 2015-12-28 | 2017-07-06 | 株式会社ブリリアントサービス | 調理用ヘッドマウントディスプレイおよび調理用ヘッドマウントディスプレイのプログラム |
| US10499997B2 (en) | 2017-01-03 | 2019-12-10 | Mako Surgical Corp. | Systems and methods for surgical navigation |
| US12383347B2 (en) | 2017-01-03 | 2025-08-12 | Mako Surgical Corp. | Systems and methods for surgical navigation |
| US11707330B2 (en) | 2017-01-03 | 2023-07-25 | Mako Surgical Corp. | Systems and methods for surgical navigation |
| US11391748B2 (en) | 2017-09-21 | 2022-07-19 | Becton, Dickinson And Company | High dynamic range assays in hazardous contaminant testing |
| US11385146B2 (en) | 2017-09-21 | 2022-07-12 | Becton, Dickinson And Company | Sampling systems and techniques to collect hazardous contaminants with high pickup and shedding efficiencies |
| US11380074B2 (en) | 2017-09-21 | 2022-07-05 | Becton, Dickinson And Company | Augmented reality devices for hazardous contaminant testing |
| US11585733B2 (en) | 2017-09-21 | 2023-02-21 | Becton, Dickinson And Company | Hazardous contaminant collection kit and rapid testing |
| US11360001B2 (en) | 2017-09-21 | 2022-06-14 | Becton, Dickinson And Company | Reactive demarcation template for hazardous contaminant testing |
| US11782042B2 (en) | 2017-09-21 | 2023-10-10 | Becton, Dickinson And Company | Hazardous contaminant collection kit and rapid testing |
| US11821819B2 (en) | 2017-09-21 | 2023-11-21 | Becton, Dickinson And Company | Demarcation template for hazardous contaminant testing |
| US12399088B2 (en) | 2017-09-21 | 2025-08-26 | Becton, Dickinson And Company | Demarcation template for hazardous contaminant testing |
| US12436158B2 (en) | 2017-09-21 | 2025-10-07 | Becton, Dickinson And Company | High dynamic range assays in hazardous contaminant testing |
| US11860173B2 (en) | 2019-01-28 | 2024-01-02 | Becton, Dickinson And Company | Hazardous contaminant collection device with integrated swab and test device |
| US11280801B2 (en) | 2019-01-28 | 2022-03-22 | Becton, Dickinson And Company | Hazardous contaminant collection device with integrated swab and test device |
Also Published As
| Publication number | Publication date |
|---|---|
| US20150347850A1 (en) | 2015-12-03 |
| EP3149564A1 (en) | 2017-04-05 |
| US9886162B2 (en) | 2018-02-06 |
| JP2017527870A (ja) | 2017-09-21 |
| JP6416290B2 (ja) | 2018-10-31 |
| CN106462322B (zh) | 2019-08-06 |
| CN106462322A (zh) | 2017-02-22 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US9886162B2 (en) | Device-provided tracking data for augmented reality | |
| US8872852B2 (en) | Positional context determination with multi marker confidence ranking | |
| CN108307214B (zh) | 用于控制装置的方法和设备 | |
| US9430877B2 (en) | Electronic device and method for selecting augmented content using the same | |
| KR102418992B1 (ko) | 전자 장치 및 전자 장치의 증강 현실 서비스 제공 방법 | |
| KR102479360B1 (ko) | 증강 현실 서비스 제공 방법 및 장치 | |
| CN103365414A (zh) | 信息处理设备、信息处理方法和信息处理系统 | |
| KR20120133648A (ko) | 증강현실 컨텐츠를 통해 타겟장치를 제어하는 제어단말기 및 서버 | |
| CN104982090A (zh) | 个人信息通信器 | |
| US20170171521A1 (en) | Projection apparatus and operation method thereof | |
| CN106131794A (zh) | 室内定位的方法及装置 | |
| CN108540542A (zh) | 一种移动增强现实系统及显示的方法 | |
| EP3229482B1 (en) | Master device, slave device, and control method therefor | |
| CN117716253A (zh) | 映射联网设备 | |
| US10652041B2 (en) | Computer vision based activation | |
| WO2017088331A1 (zh) | 控制装置及运作方法 | |
| KR101840381B1 (ko) | 다중 디바이스 연동 제어 시스템 | |
| CN107330974B (zh) | 商品展示方法、装置及移动设备 | |
| JP2012123548A5 (enExample) | ||
| CN108615260A (zh) | 一种异常现实环境下进行增强现实数字文化内容显示的方法及装置 | |
| AU2018341597B2 (en) | Using augmented reality for secure transactions | |
| US20200009791A1 (en) | Smart handling of materials for three-dimensional (3d) printers | |
| Park et al. | A Study on Intuitive IoT Interface System using 3D Depth Camera | |
| US20220417479A1 (en) | Information processing device and information processing method | |
| JP7111416B2 (ja) | 携帯端末、情報処理システム、制御方法、及びプログラム |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15728249 Country of ref document: EP Kind code of ref document: A1 |
|
| DPE1 | Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101) | ||
| REEP | Request for entry into the european phase |
Ref document number: 2015728249 Country of ref document: EP |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 2015728249 Country of ref document: EP |
|
| ENP | Entry into the national phase |
Ref document number: 2016570282 Country of ref document: JP Kind code of ref document: A |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |