WO2021061450A1 - Scene-to-text conversion - Google Patents

Scene-to-text conversion Download PDF

Info

Publication number
WO2021061450A1
WO2021061450A1 PCT/US2020/050886 US2020050886W WO2021061450A1 WO 2021061450 A1 WO2021061450 A1 WO 2021061450A1 US 2020050886 W US2020050886 W US 2020050886W WO 2021061450 A1 WO2021061450 A1 WO 2021061450A1
Authority
WO
WIPO (PCT)
Prior art keywords
audio output
implementations
user
setting
objects
Prior art date
Application number
PCT/US2020/050886
Other languages
French (fr)
Original Assignee
Qsinx Management Llc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Qsinx Management Llc filed Critical Qsinx Management Llc
Publication of WO2021061450A1 publication Critical patent/WO2021061450A1/en
Priority to US17/466,379 priority Critical patent/US20210397842A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/20Scenes; Scene-specific elements in augmented reality scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/26Techniques for post-processing, e.g. correcting the recognition result
    • G06V30/262Techniques for post-processing, e.g. correcting the recognition result using context analysis, e.g. lexical, syntactic or semantic context
    • G06V30/274Syntactic or semantic context, e.g. balancing
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • G10L13/033Voice editing, e.g. manipulating the voice of the synthesiser

Definitions

  • FIGS. 4A-4D are a flowchart representation of a method for performing scene- to-text conversion in accordance with some implementations.
  • the audio output generator 224 generates the text description based on one or more characteristic values associated with the user. For example, certain words used in the text description may be selected based on one or more of the user’s location, educational level, language proficiency, age, and/or vision capability.
  • the text generator 320 generates a text description
  • the text generator 320 generates the text description
  • the text generator 320 generates the text description
  • FIGS. 4A-4D are a flowchart representation of a method 400 for performing scene-to-text conversion in accordance with some implementations.
  • the method 400 is performed by a device (e.g., the system 200 shown in FIG. 2).
  • the method 400 is performed by processing logic, including hardware, firmware, software, or a combination thereof.
  • the method 400 is performed by a processor executing code stored in a non-transitory computer-readable medium (e.g., a memory).
  • the voice characteristic and/or the intonation characteristic of the audio output may be determined based on the location, educational level, language proficiency, age, and/or vision capability of the user.
  • the audio output may have more frequency variations when the user is a young child than when the user is an adult.
  • the characteristic value indicates a location of the user.
  • the text description may be generated based on the location of the user (e.g., a current location and/or a location of origin of the user).
  • the text generator 320 selects phrases (e.g., words) that are likely to be familiar to the user based on a location associated with the user.
  • the characteristic value indicates a language proficiency of the user.
  • the text description may be generated based on a user input indicating one or more languages in which the user is proficient and/or respective proficiency levels in each language.
  • the text generator 320 may select phrases to insert in the text description based on a language understood by the user.
  • the text generator 320 selects phrases to insert in the text description based on a level of proficiency in a language understood by the user.
  • the user input indicates a selection of a language in which the user wishes to experience the audio output, and the text generator 320 generates the text description in the selected language.
  • the language corpus provides replacement or supplemental phrases to insert in the text description.
  • the language corpus may include phrases and corresponding metadata that associates the phrases with characteristics such as a language, related phrases (e.g., “car” in English can replace or be replaced by “voiture” in French or “macchina” in Italian), grammatical gender (e.g., masculine, feminine, neuter), etc.
  • the text generator 320 may generate the text description in accordance with language-specific rules, e.g., relating to syntax.
  • the audio output is generated based on a desired narrative style of the audio output.
  • the audio output corresponds to a text description of the plurality of objects, and the text description is determined based on the desired narrative style.
  • the user may provide user input indicative of whether the user wishes to receive a nonfictional or fictional listening experience, or whether the user wishes to listen to a narrative in the style of science fiction, horror, children’s literature, etc.
  • the text generator 320 generates the text description based on the selected narrative style. For example, the text generator 320 may select phrases to insert in the text description based on the narrative style.
  • the audio output is generated based on a velocity of the user.
  • the length of the text description may be determined based on the velocity of the user.
  • the text generator 320 if the user is stationary or moving at a low velocity, the text generator 320 generates a verbose text description with a high word count. If the user is moving at a high velocity, the text generator 320 generates a terse text description with a low word count.
  • the method 400 includes synchronizing the audio output based on the location of the user.
  • the location of the user is determined using a GPS location or a wireless signal. The location may be used to pace the audio output by synchronizing the audio output with known waypoints in a physical setting.
  • the one or more communication buses 504 include circuitry that interconnects and controls communications between system components.
  • the one or more I/O devices 506 include at least one of a keyboard, a mouse, a touchpad, a joystick, one or more microphones, one or more speakers, one or more image sensors (e.g., one or more cameras, for example, a front-facing camera), one or more displays (e.g., one or more XR displays), and/or the like.
  • the memory 520 or the non-transitory computer readable storage medium of the memory 520 stores the following programs, modules and data structures, or a subset thereof including an optional operating system 530, the data obtainer 310, the text generator 320, and the speech synthesizer 330.
  • the operating system 530 includes procedures for handling various basic system services and for performing hardware dependent tasks.
  • the data obtainer 310 obtains (e.g., receives or retrieves) data, including, for example, semantic labels, an ontology, characteristic values associated with a user, a desired narrative style, a desired length of an audio output, a velocity of the user, and/or a location of the user.
  • the data obtainer 310 includes instructions 310a and heuristics and metadata 310b.
  • the text generator 320 generates a text description of the objects in the setting, e.g., based on the data obtained by the data obtainer 310.
  • the text generator 320 includes instructions 320a and heuristics and metadata 320b.
  • the speech synthesizer 330 generates an audio output based on the text description.
  • the speech synthesizer 330 includes instructions 330a and instructions and heuristics 330b.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Acoustics & Sound (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Various implementations disclosed herein include devices, systems, and methods for performing scene-to-text conversion. In various implementations, a device includes a non-transitory memory and one or more processors coupled with the non-transitory memory. In some implementations, a method includes obtaining environmental data corresponding to a setting. Based on the environmental data, a plurality of objects that are in the setting are identified. An audio output describing at least a first object of the plurality of objects in the setting is generated based on a characteristic value associated with a user of the device. The audio output is outputted.

Description

SCENE-TO-TEXT CONVERSION
CROSS-REFERENCE TO RELATED APPLICATION
[0001] This application claims the benefit of U.S. Provisional Patent App. No.
62/906,790, filed on September 27, 2019, which is incorporated by reference in its entirety.
TECHNICAL FIELD
[0002] The present disclosure generally relates to scene-to-text conversion.
BACKGROUND
[0003] Some devices are capable of generating and presenting graphical environments that include representations of physical elements. These environments may be presented on mobile communication devices.
BRIEF DESCRIPTION OF THE DRAWINGS
[0004] So that the present disclosure can be understood by those of ordinary skill in the art, a more detailed description may be had by reference to aspects of some illustrative implementations, some of which are shown in the accompanying drawings.
[0005] FIG. 1 depicts an exemplary system for use in various computer extended reality technologies, including virtual reality and mixed reality.
[0006] FIG. 2 illustrates an example system performing scene-to-text conversion according to various implementations.
[0007] FIG. 3 is a block diagram of an example audio output generator in accordance with some implementations.
[0008] FIGS. 4A-4D are a flowchart representation of a method for performing scene- to-text conversion in accordance with some implementations.
[0009] FIG. 5 is a block diagram of a device performing scene-to-text conversion in accordance with some implementations.
[0010] In accordance with common practice the various features illustrated in the drawings may not be drawn to scale. Accordingly, the dimensions of the various features may be arbitrarily expanded or reduced for clarity. In addition, some of the drawings may not depict all of the components of a given system, method or device. Finally, like reference numerals may be used to denote like features throughout the specification and figures.
SUMMARY
[0011] Various implementations disclosed herein include devices, systems, and methods for performing scene-to-text conversion. In various implementations, a device includes a non-transitory memory and one or more processors coupled with the non-transitory memory. In some implementations, a method includes obtaining environmental data corresponding to a setting. Based on the environmental data, a plurality of objects that are in the setting are identified. An audio output describing at least a first object of the plurality of objects in the setting is generated based on a characteristic value associated with a user of the device. The audio output is outputted.
[0012] In accordance with some implementations, a device includes one or more processors, a non-transitory memory, and one or more programs. In some implementations, the one or more programs are stored in the non-transitory memory and are executed by the one or more processors. In some implementations, the one or more programs include instructions for performing or causing performance of any of the methods described herein. In accordance with some implementations, a non-transitory computer readable storage medium has stored therein instructions that, when executed by one or more processors of a device, cause the device to perform or cause performance of any of the methods described herein. In accordance with some implementations, a device includes one or more processors, a non-transitory memory, and means for performing or causing performance of any of the methods described herein.
DESCRIPTION
[0013] Numerous details are described in order to provide a thorough understanding of the example implementations shown in the drawings. However, the drawings merely show some example aspects of the present disclosure and are therefore not to be considered limiting. Those of ordinary skill in the art will appreciate that other effective aspects and/or variants do not include all of the specific details described herein. Moreover, well-known systems, methods, components, devices and circuits have not been described in exhaustive detail so as not to obscure more pertinent aspects of the example implementations described herein. [0014] Some visually impaired users are not able to view settings (e.g., physical settings or computer- generated three-dimensional (3D) settings). Accordingly, a need exists for an accessibility mode that facilitates the experience of settings by visually impaired users.
[0015] The present disclosure provides methods, systems, and/or devices for performing scene-to-text conversion. In various implementations, a device includes a non- transitory memory and one or more processors coupled with the non-transitory memory. In some implementations, a method includes obtaining environmental data corresponding to a setting. Based on the environmental data, a plurality of objects that are in the setting are identified. An audio output describing at least a first object of the plurality of objects in the setting is generated based on a characteristic value associated with a user of the device. The audio output is outputted.
[0016] In various implementations, a device identifies objects in a setting and creates a narrative description for the setting. In some implementations, the device generates and presents an audio description that describes relevant portions, such as objects, in the setting. Semantic segmentation may be used to identify objects in the setting and to create the narrative description. In some implementations, characteristics of the narrative description can be configured based on, for example, characteristic values associated with a user, desired characteristics of the narrative description, and/or a velocity and/or position of the user.
[0017] Various examples of electronic systems and techniques for using such systems in relation to various extended reality technologies are described.
[0018] Physical settings are those in the world where people can sense and/or interact without use of electronic systems. For example, a room is a physical setting that includes physical elements, such as, physical chairs, physical desks, physical lamps, and so forth. A person can sense and interact with these physical elements of the physical setting through direct touch, taste, sight, smell, and hearing.
[0019] In contrast to a physical setting, an extended reality (XR) setting refers to a computer-produced environment that is partially or entirely generated using computer- produced content. While a person can interact with the XR setting using various electronic systems, this interaction utilizes various electronic sensors to monitor the person’s actions, and translates those actions into corresponding actions in the XR setting. For example, if an XR system detects that a person is looking upward, the XR system may change its graphics and audio output to present XR content in a manner consistent with the upward movement. XR settings may incorporate laws of physics to mimic physical settings.
[0020] Concepts of XR include virtual reality (VR) and augmented reality (AR).
Concepts of XR also include mixed reality (MR), which is sometimes used to refer to the spectrum of realities between physical settings (but not including physical settings) at one end and VR at the other end. Concepts of XR also include augmented virtuality (AV), in which a virtual or computer-produced setting integrates sensory inputs from a physical setting. These inputs may represent characteristics of a physical setting. For example, a virtual object may be displayed in a color captured, using an image sensor, from the physical setting. As another example, an AV setting may adopt current weather conditions of the physical setting.
[0021] Some electronic systems for implementing XR operate with an opaque display and one or more imaging sensors for capturing video and/or images of a physical setting. In some implementations, when a system captures images of a physical setting, and displays a representation of the physical setting on an opaque display using the captured images, the displayed images are called a video pass-through. Some electronic systems for implementing XR operate with an optical see-through display that may be transparent or semi-transparent (and optionally with one or more imaging sensors). Such a display allows a person to view a physical setting directly through the display, and allows for virtual content to be added to the person’s field-of-view by superimposing the content over an optical pass-through of the physical setting. Some electronic systems for implementing XR operate with a projection system that projects virtual objects onto a physical setting. The projector may present a holograph onto a physical setting, or may project imagery onto a physical surface, or may project onto the eyes (e.g., retina) of a person, for example.
[0022] Electronic systems providing XR settings can have various form factors. A smartphone or a tablet computer may incorporate imaging and display components to present an XR setting. A head-mountable system may include imaging and display components to present an XR setting. These systems may provide computing resources for generating XR settings, and may work in conjunction with one another to generate and/or present XR settings. For example, a smartphone or a tablet can connect with a head-mounted display to present XR settings. As another example, a computer may connect with home entertainment components or vehicular systems to provide an on-window display or a heads-up display. Electronic systems displaying XR settings may utilize display technologies such as LEDs, OLEDs, QD- LEDs, liquid crystal on silicon, a laser scanning light source, a digital light projector, or combinations thereof. Display technologies can employ substrates, through which light is transmitted, including light waveguides, holographic substrates, optical reflectors and combiners, or combinations thereof.
[0023] In some implementations, an electronic device comprises one or more processors working with non-transitory memory. In some implementations, the non-transitory memory stores one or more programs of executable instructions that are executed by the one or more processors. In some implementations, the executable instructions carry out the techniques and processes described herein. In some implementations, a computer (readable) storage medium has instructions that, when executed by one or more processors of an electronic device, cause the electronic device to perform, or cause performance, of any of the techniques and processes described herein. The computer (readable) storage medium is non-transitory. In some implementations, a device includes one or more processors, a non-transitory memory, and means for performing or causing performance of the techniques and processes described herein.
[0024] A head-mounted system may have one or more speaker(s) and an integrated opaque display. Alternatively, a head-mounted system may be configured to accept an external opaque display (e.g., a smartphone). The head-mounted system may incorporate one or more imaging sensors to capture images or video of the physical setting, and/or one or more microphones to capture audio of the physical setting. Rather than an opaque display, a head- mounted system may have a transparent or translucent display. The transparent or translucent display may have a medium through which light representative of images is directed to a person’s eyes. The display may utilize digital light projection, OLEDs, LEDs, uLEDs, liquid crystal on silicon, laser scanning light source, or any combination of these technologies. The medium may be an optical waveguide, a hologram medium, an optical combiner, an optical reflector, or any combination thereof. In one implementation, the transparent or translucent display may be configured to become opaque selectively. Projection-based systems may employ retinal projection technology that projects graphical images onto a person’s retina. Projection systems also may be configured to project virtual objects into the physical setting, for example, as a hologram or on a physical surface.
[0025] In accordance with some implementations, a device includes one or more processors, a non-transitory memory, and one or more programs. In some implementations, the one or more programs are stored in the non-transitory memory and are executed by the one or more processors. In some implementations, the one or more programs include instructions for perfomiing or causing performance of any of the methods described herein. In accordance with some implementations, a non-transitory computer readable storage medium has stored therein instructions that, when executed by one or more processors of a device, cause the device to perform or cause performance of any of the methods described herein. In accordance with some implementations, a device includes one or more processors, a non-transitory memory, and means for performing or causing performance of any of the methods described herein.
[0026] FIG. 1 illustrates an exemplary operating environment 100 in accordance with some implementations. While pertinent features are shown, those of ordinary skill in the art will appreciate from the present disclosure that various other features have not been illustrated for the sake of brevity and so as not to obscure more pertinent aspects of the example implementations disclosed herein. To that end, as a non-limiting example, the operating environment 100 includes an electronic device 102 and a controller 104. In some implementations, the electronic device 102 is or includes a smartphone, a tablet, a laptop computer, and/or a desktop computer. The electronic device 102 may be worn by or carried by a user 106.
[0027] As illustrated in FIG. 1, the electronic device 102 and/or the controller 104 may obtain environmental data corresponding to a setting 108. In some implementations, the setting 108 is a physical setting. The electronic device 102 and/or the controller 104 may obtain environmental data from an environmental sensor 110. In some implementations, the environmental data includes an image or a video of the physical setting, and the environmental sensor 110 may be implemented as an image sensor, e.g., a camera. In some implementations, the environmental data includes depth data, and the environmental sensor 110 may be implemented as a depth sensor.
[0028] In some implementations, the setting 108 is an extended reality (XR) setting. In some implementations, the XR setting is generated by the electronic device 102 and/or the controller 104. In some implementations, the XR setting 108 includes a virtual setting that is a simulated replacement of a physical setting. For example, the XR setting may be simulated by the electronic device 102 and/or the controller 104. In such implementations, the XR setting is different from the physical setting in which the electronic device 102 is located.
[0029] In some implementations, the XR setting includes an augmented setting that is a modified version of a physical setting. For example, in some implementations, the electronic device 102 and/or the controller 104 modify (e.g., augment) the physical setting in which the electronic device 102 is located in order to generate the XR setting. In some implementations, the electronic device 102 and/or the controller 104 generate the XR setting by simulating a replica of the physical setting in which the electronic device 102 is located. In some implementations, the electronic device 102 and/or the controller 104 generate the XR setting by removing and/or adding items from the simulated replica of the physical setting where the electronic device 102 is located.
[0030] In some implementations, the setting 108 is a computer-generated three- dimensional (3D) environment. In some implementations, the setting 108 is referred to as a graphical environment or a computer graphics environment.
[0031] In some implementations, the setting 108 includes various objects 112, such as a character object 112a, a character object 112b, a robot object 112c, and a drone object 112c (collectively referred to as objects 112). In some implementations, the objects 112 are physical objects in a physical setting. In some implementations, the objects 112 are XR objects in an XR setting. In some implementations, the objects 112 are virtual objects (e.g., AR objects). In some implementations, the objects 112 are referred to as graphical objects.
[0032] In some implementations in which the setting is an XR setting, the objects 112 are XR representations of objective-effectuators. In some implementations, the objective- effectuators model characters from fictional materials such as movies, video games, comics, and novels. For example, the character object 112a may represent and model the behavior of a character from a fictional comic, and the character object 112b represents and models the behavior of a character from a fictional video game. In some implementations, the XR setting includes XR representations of objective-effectuators that represent and model the behavior of characters from different fictional materials (e.g., from different movies/games/comics/novels). In various implementations, the objective-effectuators represent and model the behavior of physical elements (e.g., tangible objects). For example, in some implementations, the objective-effectuators model the behavior of equipment (e.g., machinery such as planes, tanks, robots, cars, etc.). In the example of FIG. 1, the robot object 112c represents and models the behavior of a robot, and the drone object 112d represents and models the behavior of a drone. In some implementations, the objective-effectuators represent and model the behavior of entities (e.g., equipment) from fictional material. In some implementations, the objective-effectuators represent and model the behavior of physical elements from the physical setting, including entities located inside and/or outside of the setting 108. In some implementations, an objective-effectuator is referred to as a virtual intelligent agent, an intelligent agent or an agent.
[0033] In various implementations, a device directs an XR representation of an objective-effectuator to perform one or more actions in order to effectuate (e.g., advance, satisfy, complete and/or achieve) one or more objectives (e.g., results and/or goals). In some implementations, the objective-effectuator is associated with a particular objective, and the XR representation of the objective-effectuator performs actions that improve the likelihood of effectuating that particular objective. In some implementations, the XR representation of the objective-effectuator corresponds to an XR affordance. In some implementations, the XR representation of the objective-effectuator is referred to as an XR object, a graphical object or a virtual object.
[0034] In some implementations, an XR representation of the objective-effectuator performs a sequence of actions. In some implementations, a device determines (e.g., generates and/or synthesizes) the actions for the objective-effectuator. In some implementations, the actions generated for the objective-effectuator are within a degree of similarity to actions that a corresponding entity (e.g., a character, an equipment and/or a thing) performs as described in fictional material or as exists in a physical setting. For example, in some implementations, an XR representation of an objective-effectuator that corresponds to a fictional action figure performs the action of flying in an XR setting because the corresponding fictional action figure flies as described in the fictional material. Similarly, in some implementations, an XR representation of an objective-effectuator that corresponds to a physical drone performs the action of hovering in an XR setting because the corresponding physical drone hovers in a physical setting. In some implementations, the device obtains the actions for the objective- effectuator. For example, in some implementations, the device receives the actions for the objective-effectuator from a separate device (e.g., a remote server) that determines the actions.
[0035] In some implementations, an objective-effectuator corresponding to a character is referred to as a character objective-effectuator, an objective of the character objective- effectuator is referred to as a character objective, and an XR representation of the character objective-effectuator is referred to as an XR character. In some implementations, the XR character performs actions in order to effectuate the character objective. In some implementations, a character objective-effectuator is referred to as a virtual character. [0036] In some implementations, an objective-effectuator corresponding to equipment
(e.g., a rope for climbing, an airplane for flying, a pair of scissors for cutting) is referred to as an equipment objective-effectuator, an objective of the equipment objective-effectuator is referred to as an equipment objective, and an XR representation of the equipment objective- effectuator is referred to as an XR equipment. In some implementations, the XR equipment performs actions in order to effectuate the equipment objective.
[0037] In some implementations, an objective-effectuator corresponding to an environment of a setting (e.g., weather pattern, features of nature and/or gravity level) is referred to as an environmental objective-effectuator, and an objective of the environmental objective-effectuator is referred to as an environmental objective. In some implementations, the environmental objective-effectuator configures an environment of the XR setting in order to effectuate the environmental objective.
[0038] In some implementations, the XR setting is generated based on a user input from the user 106. For example, in some implementations, the electronic device 102 and/or the controller 104 receive a user input indicating a terrain for the XR setting. In such implementations, the electronic device 102 and/or the controller 104 configure the XR setting such that the XR setting includes the terrain indicated via the user input. In some implementations, the user input indicates environmental conditions. In such implementations, the electronic device 102 and/or the controller 104 configure the XR setting to have the environmental conditions indicated by the user input. In some implementations, the environmental conditions include one or more of temperature, humidity, pressure, visibility, ambient light level, ambient sound level, time of day (e.g., morning, afternoon, evening, or night), and precipitation (e.g., overcast, rain, or snow).
[0039] In some implementations, the actions for the objective-effectuators are determined (e.g., generated) based on a user input from the user 106. For example, in some implementations, the mobile device receives a user input indicating placement of the objective- effectuators. In such implementations, the electronic device 102 and/or the controller 104 position the objective-effectuators in accordance with the placement indicated by the user input. In some implementations, the user input indicates specific actions that the objective- effectuators are permitted to perform. In such implementations, the electronic device 102 and/or the controller 104 select the actions for the objective-effectuator from the specific actions indicated by the user input. In some implementations, the electronic device 102 and/or the controller 104 forgo actions that are not among the specific actions indicated by the user input.
[0040] In some implementations, the setting 108 is an XR setting, and the electronic device 102 and/or the controller 104 obtain environmental data corresponding to the XR setting. The environmental data may include an image or a video of the XR setting. The environmental data may be, may include, or may be included in metadata relating to the XR setting. In some implementations, the electronic device 102 and/or the controller 104 obtain the environmental data from an XR content source 114. In some implementations, the electronic device 102 and/or the controller 104 generate the XR setting, and the electronic device 102 and/or the controller 104 obtain (e.g., synthesize) the environmental data.
[0041] In some implementations, the electronic device 102 and/or the controller 104 identify a plurality of objects that are in the setting 108 based on the environmental data. For example, in some implementations, the electronic device 102 and/or the controller 104 perform semantic segmentation on the environmental data (e.g., the image, the video, the depth data, and/or the metadata) to identify objects in the setting 108. In the example of FIG. 1, the electronic device 102 and/or the controller 104 may identify the character object 112a, the character object 112b, the robot object 112c, and the drone object 112d based on the environmental data.
[0042] In some implementations, the electronic device 102 and/or the controller 104 generate an ontology based on the semantic segmentation. The ontology may be hierarchical. For example, using semantic segmentation, the electronic device 102 and/or the controller 104 may characterize the character object 112a and the character object 112b as ‘character’ objects at a top level of the ontology and may characterize the robot object 112c and the drone object 112d as ‘equipment’ objects at a top level of the ontology. Each object 112 may have more specific ontological characteristics associated with it at deeper levels of the hierarchy. For example, the robot object 112c and the drone object 112d may be characterized as a ‘robot’ and a ‘drone,’ respectively.
[0043] In some implementations, the electronic device 102 and/or the controller 104 generate an audio output 116 describing at least one of the objects 112. The audio output 116 may be based on a characteristic (e.g., a characteristic value) associated with the user 106. For example, the audio output 116 may include a voice characteristic (e.g., accent, pitch, etc.), an intonation characteristic, and/or a text description that may be varied based on the characteristic associated with the user 106. In some implementations, for example, the text description is generated based on the user’s location, education level, age, language proficiency, etc.
[0044] In some implementations, the electronic device 102 and/or the controller 104 generate the audio output 116 based on a desired length of a listening experience. For example, if the user 106 indicates a preference for a short listening experience, the electronic device 102 and/or the controller 104 may generate a terse text description, e.g., with a low word count, for example, less than a threshold number of words. On the other hand, if the user 106 indicates a preference for a longer listening experience, the electronic device 102 and/or the controller 104 may generate a more verbose text description, e.g., with a high word count, for example, greater than a threshold number of words.
[0045] In some implementations, the electronic device 102 and/or the controller 104 generate the audio output 116 based on a velocity of the user 106. For example, if the user 106 is moving at a low velocity or is stationary, the electronic device 102 and/or the controller 104 may generate a text description with a high word count (e.g., greater than a threshold number of words). If the user 106 is moving at a high velocity, the electronic device 102 and/or the controller 104 may generate a text description with a low word count (e.g., less than a threshold number of words) or may increase the speed (e.g., words per minute) of the narration to keep pace with the movement of the user 106.
[0046] In some implementations, the electronic device 102 and/or the controller 104 generate the audio output 116 based on a position of the user 106. In some implementations, a global positioning system (GPS) location of the electronic device 102 and/or location information obtained from wireless signals may be used to identify the objects 112 and/or the setting 108. For example, semantic segmentation may be used to identify an object as a tower, while the GPS location of the electronic device 102 may be used to identify the tower as a specific tower (e.g., the Eiffel Tower).
[0047] In some implementations, the electronic device 102 and/or the controller 104 generate the audio output 116 when the electronic device 102 is in an accessibility mode. In some implementations, the electronic device 102 and/or the controller 104 generate the audio output 116 in response to determining that the user 106 is visually impaired. In some implementations, the electronic device 102 and/or the controller 104 generate the audio output 116 when the electronic device 102 is in a mode that limits display of information (e.g., in a driving mode). [0048] In some implementations, a head-mountable device (HMD) being worn by the user 106 presents (e.g., displays) the setting 108. In some implementations, the HMD includes an integrated display (e.g., a built-in display) that displays the setting 108. In some implementations, the HMD includes a head-mountable enclosure. In various implementations, the head-mountable enclosure includes an attachment region to which another device with a display can be attached. For example, in some implementations, the electronic device 102 of FIG. 1 can be attached to the head-mountable enclosure. In various implementations, the head- mountable enclosure is shaped to form a receptacle for receiving another device that includes a display (e.g., the electronic device 102). For example, in some implementations, the electronic device 102 slides or snaps into or otherwise attaches to the head-mountable enclosure. In some implementations, the display of the device attached to the head-mountable enclosure presents (e.g., displays) the setting 108. In various implementations, examples of the electronic device 102 include smartphones, tablets, media players, laptops, etc.
[0049] FIG. 2 illustrates an example system 200 that performs scene-to-text conversion according to various implementations. In some implementations, an environmental sensor 202 obtains environmental data 204 corresponding to a setting, e.g., a physical setting. For example, in some implementations, the environmental sensor 202 comprises an image sensor 206, such as a camera, that obtains an image 208 of the setting. In some implementations, the environmental sensor 202 outputs image data that the environmental sensor 202 captures, e.g., the obtained image 208.
[0050] In some implementations, the image 208 is a still image. In some implementations, the image 208 is an image frame forming part of a video feed. The image 208 includes a plurality of pixels. Some of the pixels, e.g., a first set of pixels, represent an object. Other pixels, e.g., a second set of pixels, represent a background, e.g., portions of the image 208 that do not represent the object. It will be appreciated that pixels that represent one object may represent the background for a different object.
[0051] In some implementations, the environmental sensor 202 comprises a depth sensor 210 that obtains depth data 212 corresponding to the setting. The depth data 212 may be used independently of or in connection with the image 208 to identify one or more objects in the physical setting. In some implementations, the environmental sensor 202 outputs the obtained depth data 212 corresponding to the setting. [0052] In some implementations, the setting is an XR setting generated by an XR content source 214. The XR content source 214 may provide the environmental data 204. For an XR setting, the environmental data 204 may include, for example, metadata and/or an image of the XR setting. The image may be a still image or may form a part of a video feed. The image includes a plurality of pixels. Some of the pixels, e.g., a first set of pixels, represent an object. Other pixels, e.g., a second set of pixels, represent a background, e.g., portions of the image that do not represent the object. It will be appreciated that pixels that represent one object may represent the background for a different object.
[0053] A scene-to-text conversion engine 216 receives the environmental data 204 from the environmental sensor 202 and/or the XR content source 214. In some implementations, an environmental data analyzer 218 receives the environmental data 204. In some implementations, the environmental data analyzer 218 identifies a plurality of objects that are in the setting based on the environmental data 204. For example, the environmental data analyzer 218 may perform semantic segmentation on the environmental data 204 to identify the objects in the setting. In some implementations, the environmental data 204 includes an image, and the environmental data analyzer 218 applies one or more filters and/or masks to the image to characterize pixels in the image as being associated with respective objects.
[0054] In some implementations, the environmental data analyzer 218 generates a plurality of semantic labels 220 describing the objects in the setting. In some implementations, the environmental data analyzer 218 providers the semantic labels 220 to an ontology generator 222 and/or an audio output generator 224.
[0055] In some implementations, the ontology generator 222 receives the semantic labels 220 from the environmental data analyzer 218 and generates an ontology 226 based on the semantic labels 220. In some implementations, the ontology 226 is implemented as one or more data structures representing (e.g., modeling) the objects in the setting as object entities. The ontology 226 may model relationships between the objects in the setting. For example, in some implementations, the ontology 226 is a hierarchical ontology. That is, in some implementations, the ontology 226 uses one or more hierarchical data structures to represent (e.g., model) the objects in the setting. In some implementations, for example, an object entity representing an object at one level in a hierarchy (e.g., a car or a crowd of people) may contain object entities that represent objects at a deeper level in the hierarchy (e.g., components of the car or individual members of the crowd of people). [0056] In some implementations, the audio output generator 224 receives the semantic labels 220 and/or the ontology 226 and generates an audio output 228 describing at least one object (e.g., a first object) of the plurality of objects in the setting. The audio output generator 224 may generate the audio output 228 based on the semantic labels 220 and/or the ontology 226. For example, in some implementations, the semantic labels 220 comprise a plurality of descriptors associated with the objects in the setting, and the audio output generator 224 generates a text description based on the semantic labels 220. In some implementations, the audio output generator 224 uses the ontology 226 to supplement the information conveyed by the semantic labels 220.
[0057] In some implementations, the audio output generator 224 generates the text description based on one or more characteristic values associated with the user. For example, certain words used in the text description may be selected based on one or more of the user’s location, educational level, language proficiency, age, and/or vision capability.
[0058] In some implementations, the audio output generator 224 generates the text description based on a desired narrative style, e.g., as specified by the user. For example, certain words used in the text description may be selected, e.g., from a language corpus, based on the desired narrative style.
[0059] In some implementations, the audio output generator 224 generates the audio output 228 based on a velocity of the user. The length of the text description may be determined based on the velocity of the user. For example, if the user is stationary or moving at a low velocity, the audio output generator 224 may generate a verbose text description with a high word count (e.g., greater than a threshold number of words). If the user is moving at a high velocity, the audio output generator 224 may generate a terse text description with a low word count (e.g., less than a threshold number of words). In some implementations, the audio output generator 224 selects a depth to which the ontology 226 is traversed (e.g., if the ontology 226 is hierarchical). For example, if the user is stationary or moving at a low velocity, the audio output generator 224 may traverse more levels of the ontology 226 (e.g., all levels). If the user is moving at a high velocity, the audio output generator 224 may traverse fewer levels of the ontology 226 (e.g., only the top level).
[0060] In some implementations, the audio output generator 224 performs text-to- speech conversion on the text description to generate the audio output 228. In some implementations, the audio output 228 has characteristics other than the text description itself. For example, the audio output 228 may have a voice characteristic. The voice characteristic may include, for example, the frequency and/or pitch of the voice, an accent associated with the voice, a mood associated with the voice, etc. In some implementations, the audio output 228 has an intonation characteristic. In some implementations, the audio output generator 224 determines (e.g., adjusts or selects) the voice characteristic and/or the intonation characteristic of the audio output 228 based on a characteristic value associated with the user. For example, the audio output generator 224 may determine the voice characteristic and/or the intonation characteristic of the audio output 228 based on the location, educational level, language proficiency, age, and/or vision capability of the user. In some implementations, the audio output generator 224 determines the voice characteristic and/or the intonation characteristic of the audio output 228 based on the desired narrative style of the audio output 228.
[0061] In some implementations, the audio output 228 is output via a speaker 230 that is integrated in the same electronic device as the system 200. In some implementations, the audio output 228 is provided to an accessory device 232 that is in communication with the electronic device in which the system 200 is incorporated. For example, the audio output 228 may be provided to a set of headphones or a speaker or speakers that are in wired or wireless communication with the electronic device in which the system 200 is incorporated.
[0062] FIG. 3 is a block diagram of an example audio output generator 300 in accordance with some implementations. In some implementations, the audio output generator 300 implements the audio output generator 224 shown in FIG. 2. In some implementations, the audio output generator 300 generates an audio output 302 that describes one or more objects in the setting in which a device implementing the audio output generator 300 is located. In some implementations, the audio output generator includes a data obtainer 310, a text generator 320, and a speech synthesizer 330. In various implementations, the data obtainer obtains semantic labels 304 and/or an ontology 306 that identify the one or more objects in the setting. In some implementations, the data obtainer 310 receives the semantic labels 304 from the environmental data analyzer 218 shown in FIG. 2. In some implementations, the data obtainer 310 receives the ontology 306 from the ontology generator 222 shown in FIG. 2.
[0063] In various implementations, the data obtainer 310 receives one or more characteristic values 308 that relate to a user of the device implementing the audio output generator 300. The one or more characteristic values 308 may include, for example, a location of the user (e.g., a current location and/or a location of origin of the user). In some implementations, the one or more characteristic values 308 include an education level of the user. In some implementations, the one or more characteristic values 308 include a language proficiency of the user (e.g., one or more languages in which the user is proficient and/or respective proficiency levels in each language). In some implementations, the one or more characteristic values 308 include an age of the user (e.g., a numerical age and/or an age category). In some implementations, the one or more characteristic values 308 indicate a vision capability of the user (e.g., unimpaired, vision-impaired, color-blind, blind).
[0064] In some implementations, the user 106 provides user input indicative of the one or more characteristic values 308. In some implementations, the data obtainer 310 obtains an input indicative of the one or more characteristic values from a social networking service. In some implementations, the data obtainer 310 obtains data indicative of the one or more characteristic values from a user profile.
[0065] In some implementations, the data obtainer 310 receives data indicative of a desired narrative style of the audio output 302. For example, the user may provide user input indicative of whether the user wishes to receive a nonfictional or fictional listening experience, or whether the user wishes to listen to a narrative in the style of science fiction, horror, children’s literature, etc. In some implementations, the user may provide user input indicative of a desired mood of the audio output 302.
[0066] In some implementations, the data obtainer 310 receives data indicative of the desired narrative style from the setting. For example, the data obtainer 310 may analyze the setting to estimate or predict the desired narrative style. In some implementations, the data obtainer 310 performs semantic segmentation on the setting to identify one or more objects in the setting. The data obtainer 310 may predict the desired narrative style based on the identified object or objects.
[0067] In some implementations, the data obtainer 310 receives data indicative of a desired length 314 of the audio output 302. For example, the user may provide user input that indicates that the user wishes to listen to a quick narrative. On the other hand, the user may provide user input that indicates that the user would prefer a lengthier audio experience.
[0068] In some implementations, the data obtainer 310 receives data indicative of a velocity 316 of the user. As disclosed herein, the velocity 316 may be used to pace the audio output 302 by determining the length and/or narration speed of the audio output 302 based on the velocity 316. [0069] In some implementations, the data obtainer 310 receives data indicative of a location 318 of the user. For example, the user may provide user input indicative of the user’ s current location. In some implementations, the electronic device implementing the audio output generator 300 determines the location 318, for example, using a GPS location or a wireless signal. As disclosed herein, the location 318 may be used to pace the audio output 302 by synchronizing the audio output 302 with known waypoints in a physical setting. In some implementations, the location 318 is used to identify an object in the physical setting (e.g., identifying a Ferris wheel as the High Roller based on the location 318 being indicated as Las Vegas, Nevada, or as the Singapore Flyer based on the location 318 being indicated as Singapore).
[0070] In some implementations, the text generator 320 generates a text description
322 describing the one or more objects in the setting based on the data received by the data obtainer 310. For example, the text generator 320 may generate the text description 322 based on the semantic labels 304 identifying the objects in the setting. In some implementations, the text generator 320 generates the text description 322 based on the ontology 306. For example, the ontology 306 may include information indicative of the objects in the setting. In some implementations, the ontology 306 includes information describing the objects in the setting, such as attributes (e.g., color, material, etc.) of the objects. In some implementations, the ontology 306 includes information indicative of relationships between objects in the setting. Information relating to objects in the setting may be stored at different levels of the ontology 306. In some implementations, the text generator 320 selects one or more levels of the ontology 306 to traverse in generating the text description 322, for example, based on a desired level of detail or verbosity of the text description 322.
[0071] In some implementations, the text generator 320 generates the text description
322 based on the one or more characteristic values 308 that relate to the user of the device implementing the audio output generator 300. For example, the text generator 320 may generate the text description 322 based on the location of the user (e.g., a current location and/or a location of origin of the user). In some implementations, the text generator 320 selects phrases (e.g., words) that are likely to be familiar to the user based on a location associated with the user. For example, if an object in the setting is associated with the semantic label ‘truck’ and the user is based in the United Kingdom, the text generator 320 may use the phrase ‘lorry’ to describe the object associated with the semantic label ‘truck.’ In some implementations, the text generator 320 uses a language corpus 324 to obtain supplemental or replacement phrases to insert in the text description 322. The language corpus 324 may be stored in a memory and may include phrases and corresponding metadata. In some implementations, the metadata associates the phrases with characteristics such as formality level (e.g., formal vs. casual), geographic area (e.g., United Kingdom, Southern United States, Midwestern United States), related phrases (e.g., ‘flat’ is synonymous with and can replace ‘apartment’), etc.
[0072] In some implementations, the text generator 320 generates the text description
322 based on an education level of the user. For example, the text generator 320 may select phrases to insert in the text description 322 based on a reading level of the user. In some implementations, the language corpus 324 provides supplemental or replacement phrases to insert in the text description 322. In some implementations, the text generator 320 determines an organization of the text description 322, e.g., sentence structures and/or paragraph structures, based on the education level of the user. For example, for a user with a high school diploma, the text generator 320 may generate a text description with simple sentence and paragraph structures. The text generator 320 may generate a text description with complex sentence and paragraph structures for a user with a college degree in English literature. The text generator 320 may obtain rules 326 governing the text description 322 and may generate the text description 322 in accordance with the rules 326.
[0073] In some implementations, the text generator 320 generates the text description
322 based on a language proficiency of the user (e.g., one or more languages in which the user is proficient and/or respective proficiency levels in each language). For example, the text generator 320 may select phrases to insert in the text description 322 based on a language understood by the user. In some implementations, the text generator 320 selects phrases to insert in the text description 322 based on a level of proficiency in a language understood by the user. In some implementations, e.g., if the user is proficient in multiple languages, the data obtainer 310 receives a user input indicating a selection of a preferred language in which the user wishes to experience the audio output 302, and the text generator 320 generates the text description 322 in the selected language. In some implementations, the language corpus 324 provides replacement or supplemental phrases to insert in the text description 322. The language corpus 324 may be stored in a memory and may include phrases and corresponding metadata. In some implementations, the metadata associates the phrases with characteristics such as a language, related phrases (e.g., ‘house’ in English can replace or be replaced by ‘maison’ in French or ‘casa’ in Spanish), grammatical gender (e.g., masculine, feminine, neuter), etc. In some implementations, the text generator 320 obtains rules 326, e.g., grammatical rules, and generates the text description 322 in accordance with the rules 326.
[0074] In some implementations, the text generator 320 generates the text description
322 based on an age of the user (e.g., a numerical age and/or an age category). For example, the text generator 320 may select phrases to insert in the text description 322 based on an age category of the user. In some implementations, the language corpus 324 provides replacement or supplemental phrases to insert in the text description 322. The language corpus 324 may include phrases and corresponding metadata. In some implementations, the metadata associates phrases with related phrases (e.g., ‘dog’ can replace or be replaced by ‘doggy’ for a younger user). In some implementations, the rules 326 specify phrases that are to be omitted from the audio output 302 based on the user’s age. For example, a rule 326 may specify that if the characteristic values 308 indicate that the user is below a threshold age, certain phrases (e.g., profanity or other potentially offensive phrases) be omitted from the text description 322.
[0075] In some implementations, the text generator 320 generates the text description
322 based on a vision capability of the user (e.g., unimpaired, vision-impaired, color-blind, or blind). For example, the text generator 320 may select phrases to insert in the text description 322 based on the vision capability of the user. In some implementations, the rules 326 specify phrases that are to be omitted from the audio output 302 based on the user’s vision capability. For example, a rule 326 may specify that if the characteristic values 308 indicate that the user is blind and may not have an understanding of visual concepts, phrases indicative of such visual concepts may be omitted from the text description 322.
[0076] In some implementations, the text generator 320 generates the text description
322 based on the narrative style 312. For example, the text generator 320 may select phrases to insert in the text description 322 based on the narrative style 312. In some implementations, the text generator 320 selects phrases (e.g., words) that are associated with the narrative style. For example, if an object in the setting is associated with the semantic label ‘animal,’ the text generator 320 may use the phrase ‘animal’ to describe the object in a nonfiction narrative style. For a science fiction narrative style, the text generator 320 may use the phrase ‘lifeform’ to describe the same object. For a horror fiction narrative style, the text generator 320 may use the phrase ‘creature’ or ‘monster’ to describe the same object. In some implementations, the text generator 320 uses a language corpus 324 to obtain supplemental or replacement phrases to insert in the text description 322. The language corpus 324 may be stored in a memory and may include phrases and corresponding metadata. In some implementations, the metadata associates the phrases with characteristics such as associated narrative styles and related phrases (e.g., ‘animal’ is synonymous with and can replace or be replaced by ‘lifeform,’ ‘creature,’ or ‘monster’), etc.
[0077] In some implementations, the text generator 320 generates the text description
322 based on a velocity of the user. The length of the text description may be determined based on the velocity of the user. For example, if the user is stationary or moving at a low velocity (e.g., at less than a threshold velocity), the text generator 320 may generate a text description 322 with a high word count (e.g., greater than a threshold number of words). If the user is moving at a high velocity (e.g., at greater than a threshold velocity), the text generator 320 may generate a text description 322 with a low word count (e.g., less than a threshold number of words). In some implementations, the text generator 320 selects a depth to which the ontology 306 is traversed (e.g., if the ontology 306 is hierarchical). For example, if the user is stationary or moving at a low velocity, the text generator 320 may traverse more levels of the ontology 306 (e.g., all levels). If the user is moving at a high velocity, the text generator 320 may traverse fewer levels of the ontology 306 (e.g., only the top level).
[0078] In some implementations, the speech synthesizer 330 performs text-to-speech conversion on the text description 322 to generate the audio output 302. In some implementations, the audio output 302 has characteristics other than the text description itself. For example, the audio output 302 may have a voice characteristic. The voice characteristic may include, for example, the frequency and/or pitch of the voice, an accent associated with the voice, a mood associated with the voice, etc. In some implementations, the audio output 302 has an intonation characteristic. In some implementations, the speech synthesizer 330 determines (e.g., adjusts or selects) the voice characteristic and/or the intonation characteristic of the audio output 302 based on a characteristic value associated with the user. For example, the speech synthesizer 330 may determine the voice characteristic and/or the intonation characteristic of the audio output 302 based on the location, educational level, language proficiency, age, and/or vision capability of the user. In some implementations, the speech synthesizer 330 determines the voice characteristic and/or the intonation characteristic of the audio output 302 based on the desired narrative style of the audio output 302.
[0079] In some implementations, the audio output 302 is output via a speaker integrated in the electronic device implementing the audio output generator 300. In some implementations, the audio output 302 is provided to an accessory device that is in communication with the electronic device implementing the audio output generator. For example, the audio output 302 may be provided to a set of headphones or a speaker or speakers via a wired or wireless communication link.
[0080] FIGS. 4A-4D are a flowchart representation of a method 400 for performing scene-to-text conversion in accordance with some implementations. In various implementations, the method 400 is performed by a device (e.g., the system 200 shown in FIG. 2). In some implementations, the method 400 is performed by processing logic, including hardware, firmware, software, or a combination thereof. In some implementations, the method 400 is performed by a processor executing code stored in a non-transitory computer-readable medium (e.g., a memory). Briefly, in various implementations, the method 400 includes obtaining environmental data corresponding to a setting, identifying objects in the setting based on the environmental data, generating an audio output describing one or more objects in the setting based on a characteristic value associated with a user of the device, and outputting the audio output.
[0081] Referring to FIG. 4A, as represented by block 410, the method 400 may include obtaining environmental data corresponding to a setting. In some implementations, the setting is a physical setting. Referring now to FIG. 4B, as represented by block 410a, the environmental data may include an image of the setting. In some implementations, the image is a still frame. In some implementations, the image is an image frame forming part of a video feed. The image of the setting may be received from an image sensor (e.g., a camera), as represented by block 410b. In some implementations, the setting is an XR setting, and the image of the setting is received from an XR content source. As represented by block 410c, in some implementations, the environmental data includes depth data. The depth data may be received from a depth sensor, as represented by block 410d.
[0082] As represented by block 420, in some implementations, the method 400 includes identifying a plurality of objects in the setting based on the environmental data. As represented by block 420a, in some implementations, semantic segmentation may be performed to identify the objects in the setting. For example, in some implementations, the environmental data includes an image, and one or more filters and/or masks are applied to the image to characterize pixels in the image as being associated with respective objects. Semantic segmentation is then performed to generate semantic labels corresponding to the objects represented by the pixels.
[0083] In some implementations, as represented by block 420b, an ontology is generated based on the semantic segmentation. For example, in some implementations, one or more data structures may be generated to represent (e.g., model) the objects in the setting as object entities. The ontology may model relationships between the objects in the setting. In some implementations, as represented by block 420c, the ontology is hierarchical. For example, one or more data structures may be implemented as a hierarchical data structure, such as a node graph. In some implementations, for example, a top-level node represents an object at one level in a hierarchy (e.g., a car or a crowd of people). Child nodes in the node graph may represent objects at a deeper level in the hierarchy (e.g., components of the car or individual members of the crowd of people). In some implementations, child nodes represent attributes or characteristics of the objects represented by their respective parent nodes.
[0084] As represented by block 430, in some implementations, the method 400 includes generating an audio output based on a characteristic value associated with a user of the device. The audio output describes at least a first object of the plurality of objects in the setting. Referring now to FIG. 4C, in some implementations, the audio output comprises a voice characteristic that is determined based on the characteristic value, as represented by block 430a. The voice characteristic may include, for example, the frequency and/or pitch of the voice, an accent associated with the voice, a mood associated with the voice, etc. In some implementations, as represented by block 430b, the audio output comprises an intonation characteristic that is determined based on the characteristic value. For example, the voice characteristic and/or the intonation characteristic of the audio output may be determined based on the location, educational level, language proficiency, age, and/or vision capability of the user. In some implementations, for example, the audio output may have more frequency variations when the user is a young child than when the user is an adult.
[0085] In some implementations, as represented by block 430c, the audio output corresponds to a text description of the plurality of objects, and the method 400 includes determining the text description based on the characteristic value. For example, phrases may be selected for inclusion in the text description from a language corpus stored in a memory. In some implementations, the language corpus includes phrases and corresponding metadata. In some implementations, the metadata associates the phrases with characteristics such as formality level (e.g., formal vs. casual), geographic area (e.g., United Kingdom, Southern United States, Midwestern United States), related phrases (e.g., “flat” is synonymous with and can replace “apartment”), grammatical gender, etc. In some implementations, the selection of phrases for inclusion in the text description is guided by rules that may specify, for example, sentence and paragraph structures, phrases to avoid using, and/or other constraints on the text description.
[0086] In some implementations, as represented by block 430d, the characteristic value indicates a location of the user. For example, the text description may be generated based on the location of the user (e.g., a current location and/or a location of origin of the user). In some implementations, the text generator 320 selects phrases (e.g., words) that are likely to be familiar to the user based on a location associated with the user. For example, if an object in the setting is associated with the semantic label “truck” and the user is based in the United Kingdom, the text generator 320 may use the phrase “lorry” to describe the object associated with the semantic label “truck.” In some implementations, the current location of the user is used to identify an object in the physical setting as a particular object (e.g., identifying a Ferris wheel as the High Roller based on the location being indicated as Las Vegas, Nevada, or as the Singapore Flyer based on the location being indicated as Singapore).
[0087] In some implementations, as represented by block 430e, the characteristic value indicates an educational level of the user. For example, the text generator 320 may select phrases to insert in the text description based on a reading level of the user. In some implementations, the language corpus provides supplemental or replacement phrases to insert in the text description. In some implementations, the text generator 320 determines an organization of the text description, e.g., sentence structures and/or paragraph structures, based on the education level of the user. For example, for a user with a high school diploma, the text generator 320 may generate a text description with simple sentence and paragraph structures. The text generator 320 may generate a text description with complex sentence and paragraph structures for a user with a college degree in English literature.
[0088] In some implementations, as represented by block 430f, the characteristic value indicates a language proficiency of the user. For example, the text description may be generated based on a user input indicating one or more languages in which the user is proficient and/or respective proficiency levels in each language. For example, the text generator 320 may select phrases to insert in the text description based on a language understood by the user. In some implementations, the text generator 320 selects phrases to insert in the text description based on a level of proficiency in a language understood by the user. In some implementations, e.g., if the user is proficient in multiple languages, the user input indicates a selection of a language in which the user wishes to experience the audio output, and the text generator 320 generates the text description in the selected language. In some implementations, the language corpus provides replacement or supplemental phrases to insert in the text description. The language corpus may include phrases and corresponding metadata that associates the phrases with characteristics such as a language, related phrases (e.g., “car” in English can replace or be replaced by “voiture” in French or “macchina” in Italian), grammatical gender (e.g., masculine, feminine, neuter), etc. The text generator 320 may generate the text description in accordance with language-specific rules, e.g., relating to syntax.
[0089] In some implementations, as represented by block 430g, the characteristic value indicates an age of the user (e.g., a numerical age and/or an age category). For example, the text generator 320 may select phrases to insert in the text description based on an age category of the user. In some implementations, the language corpus provides replacement or supplemental phrases to insert in the text description. The language corpus may include phrases and corresponding metadata. In some implementations, the metadata associates phrases with related phrases (e.g., “cat” can replace or be replaced by “kitty” for a younger user). In some implementations, rules specify phrases that are to be omitted from the audio output based on the user’ s age. For example, a rule may specify that if the characteristic value indicates that the user is a young child, certain phrases (e.g., profanity or other potentially offensive phrases) be omitted from the text description.
[0090] In some implementations, as represented by block 430h, the characteristic value indicates a vision capability of the user. The text generator 320 may generate the text description based on the vision capability of the user. For example, the text generator 320 may select phrases to insert in the text description based on the vision capability of the user. In some implementations, the rules specify phrases that are to be omitted from the audio output based on the user’ s vision capability.
[0091] In some implementations, as represented by block 430i, the method includes generating the audio output based on a desired length of the audio output. As represented by block 430j, in some implementations, the audio output corresponds to a text description that is determined based on the desired length of the audio output. For example, if the user indicates a preference for a short listening experience, the text generator 320 may generate the text description with a low word count. On the other hand, if the user indicates a preference for a longer listening experience, the text generator 320 may generate a more verbose text description, e.g., with a high word count. In some implementations, as represented by block 430k, an ontology is generated based on the semantic segmentation, and the text generator 320 selects a depth of the ontology to traverse in generating the text description based on the desired length of the audio output. The audio output is generated based on the hierarchy, e.g., the traversed portion of the hierarchy.
[0092] Referring to FIG. 4D, in some implementations, as represented by block 4301, the audio output is generated based on a desired narrative style of the audio output. For example, as represented by block 430m, the audio output corresponds to a text description of the plurality of objects, and the text description is determined based on the desired narrative style. In some implementations, the user may provide user input indicative of whether the user wishes to receive a nonfictional or fictional listening experience, or whether the user wishes to listen to a narrative in the style of science fiction, horror, children’s literature, etc. In some implementations, the text generator 320 generates the text description based on the selected narrative style. For example, the text generator 320 may select phrases to insert in the text description based on the narrative style. In some implementations, the text generator 320 selects phrases (e.g., words) that are associated with the narrative style. For example, if an object in the setting is associated with the semantic label “person,” the text generator 320 may use the phrase “person” to describe the object in a nonfiction narrative style. For a science fiction narrative style, the text generator 320 may use the phrase “Earthling” to describe the same object. In some implementations, the text generator 320 uses a language corpus to obtain supplemental or replacement phrases to insert in the text description. The language corpus may include phrases and corresponding metadata that associates the phrases with characteristics such as associated narrative styles and related phrases (e.g., “person” is synonymous with and can replace or be replaced by “human,” “Earthling”), etc.
[0093] In some implementations, as represented by block 430n, the audio output comprises a voice characteristic that is determined based on the desired narrative style. For example, if the user selects a science fiction narrative style, the audio output may be characterized by a robotic voice. In some implementations, as represented by block 430o, the audio output comprises an intonation characteristic that is determined based on the desired narrative style. For example, if the user selects a nonfiction narrative style, the audio output may be characterized by a monotonous voice.
[0094] In some implementations, as represented by block 430p, the audio output is generated based on a velocity of the user. For example, the length of the text description may be determined based on the velocity of the user. In some implementations, if the user is stationary or moving at a low velocity, the text generator 320 generates a verbose text description with a high word count. If the user is moving at a high velocity, the text generator 320 generates a terse text description with a low word count.
[0095] In some implementations, as represented by block 430q, an ontology is generated based on the semantic segmentation, and the text generator 320 selects a depth of the ontology to traverse in generating the text description based on the velocity of the user. For example, if the user is stationary or moving at a low velocity, the text generator 320 may traverse more levels of the ontology (e.g., all levels). If the user is moving at a high velocity, the text generator 320 may traverse fewer levels of the ontology (e.g., only the top level). The audio output is generated based on the hierarchy, e.g., the traversed portion of the hierarchy.
[0096] In some implementations, as represented by block 43 Or, the method 400 includes synchronizing the audio output based on the location of the user. For example, in some implementations, as represented by block 430s, the location of the user is determined using a GPS location or a wireless signal. The location may be used to pace the audio output by synchronizing the audio output with known waypoints in a physical setting.
[0097] In some implementations, as represented by block 440, the method 400 includes outputting the audio output. In some implementations, as represented by block 440a, the audio output is output via a speaker that is integrated in the device that performs the scene-to-text conversion. In some implementations, as represented by block 440b, the audio output is provided to an accessory device that is in communication with the device that performs the scene-to-text conversion. For example, the audio output may be provided to a set of headphones or a speaker or speakers that are in wired or wireless communication with the device that performs the scene-to-text conversion.
[0098] In some implementations, the method 400 includes generating an ontology for the setting based on a semantic segmentation of the environmental data. In some implementations, the ontology indicates a hierarchy of the plurality of objects. In some implementations, the method 400 includes ordering respective descriptions of the plurality of objects in the audio output based on the hierarchy indicated by the ontology. As such, in various implementations, the device outputs an audio description of an object that is towards a top of the hierarchy before outputting an audio description of an object that is towards a bottom of the hierarchy. Since objects towards the top of the hierarchy may be more relevant than objects towards the bottom of the hierarchy, the user hears audio descriptions of objects that may be more relevant before hearing audio descriptions of objects that may be less relevant. [0099] FIG. 5 is a block diagram of a device 500 that performs scene-to-text conversion in accordance with some implementations. While certain specific features are illustrated, those skilled in the art will appreciate from the present disclosure that various other features have not been illustrated for the sake of brevity, and so as not to obscure more pertinent aspects of the implementations disclosed herein. To that end, as a non-limiting example, in some implementations, the device 500 includes one or more processing units 502 (e.g., microprocessors, application- specific integrated-circuits (ASICs), field-programmable gate arrays (FPGAs), graphics processing units (GPUs), central processing units (CPUs), processing cores, and/or the like), one or more input/output (EO) devices 506, one or more communication interfaces 508 (e.g., universal serial bus (USB), FIREWIRE, THUNDERBOLT, IEEE 802.3x, IEEE 802.1 lx, IEEE 802.16x, global system for mobile communications (GSM), code division multiple access (CDMA), time division multiple access (TDMA), global positioning system (GPS), infrared (IR), BLUETOOTH, ZIGBEE, and/or the like type interface), one or more programming (e.g., I/O) interfaces 510, a memory 520, and one or more communication buses 504 for interconnecting these and various other components.
[00100] In some implementations, the one or more communication buses 504 include circuitry that interconnects and controls communications between system components. In some implementations, the one or more I/O devices 506 include at least one of a keyboard, a mouse, a touchpad, a joystick, one or more microphones, one or more speakers, one or more image sensors (e.g., one or more cameras, for example, a front-facing camera), one or more displays (e.g., one or more XR displays), and/or the like.
[00101] The memory 520 may include high-speed random-access memory, such as dynamic random-access memory (DRAM), static random-access memory (ERAM), double - data-rate random-access memory (DDR RAM), or other random-access solid-state memory devices. In some implementations, the memory 520 includes non-volatile memory, such as one or more magnetic disk storage devices, optical disk storage devices, flash memory devices, or other non-volatile solid-state storage devices. The memory 520 optionally includes one or more storage devices remotely located from the one or more processing units 502. The memory 520 comprises a non-transitory computer readable storage medium. In some implementations, the memory 520 or the non-transitory computer readable storage medium of the memory 520 stores the following programs, modules and data structures, or a subset thereof including an optional operating system 530, the data obtainer 310, the text generator 320, and the speech synthesizer 330. [00102] The operating system 530 includes procedures for handling various basic system services and for performing hardware dependent tasks.
[00103] As described herein, in various implementations, the data obtainer 310 obtains (e.g., receives or retrieves) data, including, for example, semantic labels, an ontology, characteristic values associated with a user, a desired narrative style, a desired length of an audio output, a velocity of the user, and/or a location of the user. To that end, the data obtainer 310 includes instructions 310a and heuristics and metadata 310b. As described herein, in various implementations, the text generator 320 generates a text description of the objects in the setting, e.g., based on the data obtained by the data obtainer 310. To that end, the text generator 320 includes instructions 320a and heuristics and metadata 320b. As described herein, in various implementations, the speech synthesizer 330 generates an audio output based on the text description. To that end, the speech synthesizer 330 includes instructions 330a and instructions and heuristics 330b.
[00104] Although the data obtainer 310, the text generator 320, and the speech synthesizer 330 are shown as residing on a single device, it should be understood that in other implementations, any combination of the data obtainer 310, the text generator 320, and the speech synthesizer 330 may be located in separate computing devices.
[00105] Moreover, FIG. 5 is intended more as functional description of the various features which be present in a particular implementation as opposed to a structural schematic of the implementations described herein. As recognized by those of ordinary skill in the art, items shown separately could be combined and some items could be separated. For example, some functional blocks shown separately in FIG. 5 could be implemented as a single block and the various functions of single functional blocks could be implemented by one or more functional blocks in various implementations. The actual number of blocks and the division of particular functions and how features are allocated among them will vary from one implementation to another and, in some implementations, depends in part on the particular combination of hardware, software, and/or firmware chosen for a particular implementation.
[00106] Various techniques discussed herein consider the use of location information to affect visual characteristics. Implemented of such technologies are reminded to inform and obtain a user’s permission for using such information. Location information should be utilized by appropriate parties only for legitimate and reasonable purposes. Those parties utilizing such information will adhere to privacy policies and practices that are at least in accordance with appropriate laws and regulations. In addition, such policies are to be well-established, user- accessible, and recognized as in compliance with or above governmental and industry standards.
[00107] It is possible to implement the techniques described herein under an opt-in model, such that a user’s informed consent is obtained before their location is used to provide the features contemplated. Settings or other preferences may be adjusted such that users can decide whether to enable the techniques described herein. It is also possible to limit the degree to which location information is obtained and/or used. For instance, locations can be obscured such that, for example, the location information identifies the country but does not identify a precise location of the user.
[00108] While various aspects of implementations within the scope of the appended claims are described above, it should be apparent that the various features of implementations described above may be embodied in a wide variety of forms and that any specific structure and/or function described above is merely illustrative. Based on the present disclosure one skilled in the art should appreciate that an aspect described herein may be implemented independently of any other aspects and that two or more of these aspects may be combined in various ways. For example, an apparatus may be implemented and/or a method may be practiced using any number of the aspects set forth herein. In addition, such an apparatus may be implemented and/or such a method may be practiced using other structure and/or functionality in addition to or other than one or more of the aspects set forth herein.

Claims

What is claimed is:
1. A method comprising: at a device including a non-transitory memory and one or more processors coupled with the non-transitory memory: obtaining environmental data corresponding to a setting; identifying, based on the environmental data, a plurality of objects that are in the setting; generating, based on a characteristic value associated with a user of the device, an audio output describing at least a first object of the plurality of objects in the setting; and outputting the audio output.
2. The method of claim 1, wherein the environmental data comprises an image of the setting.
3. The method of claim 2, further comprising obtaining the image of the setting using an image sensor of the device.
4. The method of any of claims 1-3, wherein the environmental data comprises depth data.
5. The method of claim 4, further comprising obtaining the depth data using a depth sensor of the device.
6. The method of any of claims 1-5, further comprising: performing semantic segmentation on the environmental data; and identifying the plurality of objects based on the semantic segmentation.
7. The method of claim 6, further comprising generating an ontology based on the semantic segmentation.
8. The method of claim 7, wherein the ontology is hierarchical.
9. The method of any of claims 1-8, wherein the audio output comprises a voice characteristic, the method further comprising determining the voice characteristic based on the characteristic value.
10. The method of any of claims 1-9, wherein the audio output comprises an intonation characteristic, the method further comprising determining the intonation characteristic based on the characteristic value.
11. The method of any of claims 1-10, wherein the audio output corresponds to a text description of the plurality of objects, the method further comprising determining the text description based on the characteristic value.
12. The method of claim 11, wherein the characteristic value indicates a location of the user.
13. The method of any of claims 11-12, wherein the characteristic value indicates an educational level of the user.
14. The method of any of claims 11-13, wherein the characteristic value indicates a language preference of the user.
15. The method of any of claims 11-14, wherein the characteristic value indicates an age of the user.
16. The method of any of claims 11-15, wherein the characteristic value indicates a vision capability of the user.
17. The method of any of claims 1-16, further comprising generating the audio output based on a desired length of the audio output.
18. The method of claim 17, wherein the audio output corresponds to a text description of the plurality of objects, the method further comprising determining the text description based on the desired length of the audio output.
19. The method of claim 18, further comprising: generating an ontology based on a semantic segmentation of the environmental data, wherein the ontology has a hierarchy; selecting a depth of the hierarchy to traverse based on the desired length of the audio output; and generating the audio output based on the hierarchy.
20. The method of any of claims 1-19, further comprising generating the audio output based on a desired narrative style of the audio output.
21. The method of claim 20, wherein the audio output corresponds to a text description of the plurality of objects, the method further comprising determining the text description based on the desired narrative style of the audio output.
22. The method of any of claims 20-21, wherein the audio output comprises a voice characteristic, the method further comprising determining the voice characteristic based on the desired narrative style of the audio output.
23. The method of any of claims 20-22, wherein the audio output comprises an intonation characteristic, the method further comprising determining the intonation characteristic based on the desired narrative style of the audio output.
24. The method of any of claims 1-23, further comprising generating the audio output based on a velocity of the user.
25. The method of claim 24, further comprising: generating an ontology based on the semantic segmentation, wherein the ontology has a hierarchy; selecting a depth of the hierarchy to traverse based on the velocity of the user; and generating the audio output based on the hierarchy.
26. The method of any of claims 1-25, further comprising synchronizing the audio output based on a location of the user.
27. The method of claim 26, further comprising obtaining the location of the user using at least one of a global positioning system (GPS) location or a wireless signal.
28. The method of any of claims 1-27, wherein the device comprises a head-mountable device (HMD).
29. The method of any of claims 1-28, wherein the device comprises a speaker coupled with the one or more processors.
30. The method of claim 29, further comprising outputting the audio output via the speaker.
31. The method of any of claims 1-30, further comprising outputting the audio output via an accessory device in communication with the device.
32. The method of any of claims 1-31, wherein the setting comprises a virtual reality (VR) setting.
33. The method of any of claims 1-32, wherein the setting comprises a mixed reality (MR) setting.
34. The method of any of claims 1-33, further comprising: generating an ontology for the setting based on a semantic segmentation of the environmental data, wherein the ontology indicates a hierarchy of the plurality of objects; and ordering respective descriptions of the plurality of objects in the audio output based on the hierarchy indicated by the ontology.
35. A device comprising: one or more processors; a non- transitory memory; and one or more programs stored in the non-transitory memory, which, when executed by the one or more processors, cause the device to perform any of the methods of claims 1-34.
36. A non-transitory memory storing one or more programs, which, when executed by one or more processors of a device, cause the device to perform any of the methods of claims 1-34.
37. A device comprising: one or more processors; a non- transitory memory; and means for causing the device to perform any of the methods of claims 1-34.
PCT/US2020/050886 2019-09-27 2020-09-15 Scene-to-text conversion WO2021061450A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/466,379 US20210397842A1 (en) 2019-09-27 2021-09-03 Scene-to-Text Conversion

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201962906790P 2019-09-27 2019-09-27
US62/906,790 2019-09-27

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/466,379 Continuation US20210397842A1 (en) 2019-09-27 2021-09-03 Scene-to-Text Conversion

Publications (1)

Publication Number Publication Date
WO2021061450A1 true WO2021061450A1 (en) 2021-04-01

Family

ID=72659948

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2020/050886 WO2021061450A1 (en) 2019-09-27 2020-09-15 Scene-to-text conversion

Country Status (2)

Country Link
US (1) US20210397842A1 (en)
WO (1) WO2021061450A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102349355B1 (en) * 2021-10-27 2022-01-10 주식회사 안심엘피씨 Method, device and system for constructing deboning education contents based on ar/vr environment using motion gloves

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006030464A (en) * 2004-07-14 2006-02-02 Nec Corp Audio output device, character image display device, audio output method and character image display method
US9792501B1 (en) * 2016-12-31 2017-10-17 Vasuyantra Corp. Method and device for visually impaired assistance

Family Cites Families (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2378301A (en) * 2001-07-31 2003-02-05 Hewlett Packard Co Personal object recognition system for visually impaired persons
US7495631B2 (en) * 2005-04-12 2009-02-24 International Business Machines Corporation Method, apparatus and computer program product for dynamic display of billboard information
US8068644B2 (en) * 2006-03-07 2011-11-29 Peter Thomas Tkacik System for seeing using auditory feedback
WO2008118441A1 (en) * 2007-03-26 2008-10-02 Mix & Burn, Llc Systems and methods for enabling users to sample and acquire content
US9226104B2 (en) * 2008-11-03 2015-12-29 Sprint Spectrum L.P. Methods and systems for disabling text messaging while driving
CN102907109A (en) * 2011-03-18 2013-01-30 松下电器产业株式会社 Glasses, stereoscopic image processing device, and system
WO2014043543A1 (en) * 2012-09-13 2014-03-20 Parametric Sound Corporation Personal audio system and method
CN105213157A (en) * 2014-06-13 2016-01-06 阿里巴巴集团控股有限公司 A kind of blind-guiding method and system
US10390059B2 (en) * 2014-09-04 2019-08-20 Comcast Cable Communications, Llc Latent binding of content based on user preference
WO2016172305A1 (en) * 2015-04-21 2016-10-27 Freedom Scientific, Inc. Method and system for converting text to speech
CN206214373U (en) * 2016-03-07 2017-06-06 维看公司 Object detection from visual information to blind person, analysis and prompt system for providing
US10782780B2 (en) * 2016-12-31 2020-09-22 Vasuyantra Corp. Remote perception of depth and shape of objects and surfaces
US10528815B2 (en) * 2016-12-31 2020-01-07 Vasuyantra Corp. Method and device for visually impaired assistance
JP6617783B2 (en) * 2018-03-14 2019-12-11 カシオ計算機株式会社 Information processing method, electronic device, and program
US11025815B2 (en) * 2018-09-29 2021-06-01 Apple Inc. Devices, methods, and graphical user interfaces for assisted photo-taking
US10726062B2 (en) * 2018-11-30 2020-07-28 Sony Interactive Entertainment Inc. System and method for converting image data into a natural language description
US11455796B1 (en) * 2019-07-23 2022-09-27 Snap Inc. Blindness assist glasses
US11551688B1 (en) * 2019-08-15 2023-01-10 Snap Inc. Wearable speech input-based vision to audio interpreter
EP3918451B1 (en) * 2019-09-27 2023-11-29 Apple Inc. Content generation based on audience engagement
US11729476B2 (en) * 2021-02-08 2023-08-15 Sony Group Corporation Reproduction control of scene description

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006030464A (en) * 2004-07-14 2006-02-02 Nec Corp Audio output device, character image display device, audio output method and character image display method
US9792501B1 (en) * 2016-12-31 2017-10-17 Vasuyantra Corp. Method and device for visually impaired assistance

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
JUAN DIEGO GOMEZ ET AL: "Toward 3D scene understanding via audio-description", COMPUTERS AND ACCESSIBILITY, ACM, 2 PENN PLAZA, SUITE 701 NEW YORK NY 10121-0701 USA, 24 October 2011 (2011-10-24), pages 293 - 294, XP058005633, ISBN: 978-1-4503-0920-2, DOI: 10.1145/2049536.2049613 *
ZAND MOHSEN ET AL: "Ontology-Based Semantic Image Segmentation Using Mixture Models and Multiple CRFs", IEEE TRANSACTIONS ON IMAGE PROCESSING, IEE SERVICE CENTER , PISCATAWAY , NJ, US, vol. 25, no. 7, 8 April 2016 (2016-04-08), pages 3233 - 3248, XP011611084, ISSN: 1057-7149, [retrieved on 20160519], DOI: 10.1109/TIP.2016.2552401 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102349355B1 (en) * 2021-10-27 2022-01-10 주식회사 안심엘피씨 Method, device and system for constructing deboning education contents based on ar/vr environment using motion gloves

Also Published As

Publication number Publication date
US20210397842A1 (en) 2021-12-23

Similar Documents

Publication Publication Date Title
US11798242B2 (en) Contextual computer-generated reality (CGR) digital assistants
US10922049B2 (en) Natural language based computer animation
TWI581178B (en) User controlled real object disappearance in a mixed reality display
US20200034025A1 (en) Systems and methods for multisensory semiotic communications
US20210398360A1 (en) Generating Content Based on State Information
US20210365107A1 (en) Object positioning and movement in three dimensional content
US20210397842A1 (en) Scene-to-Text Conversion
US20230377237A1 (en) Influencing actions of agents
US10891922B1 (en) Attention diversion control
US20220262081A1 (en) Planner for an objective-effectuator
WO2020219643A1 (en) Training a model with human-intuitive inputs
KR102484333B1 (en) Creation of Objectives for Objective Executors in Synthetic Reality Settings
KR102499247B1 (en) Goal-executors in synthetic reality settings
US11055930B1 (en) Generating directives for objective-effectuators
US11393135B1 (en) Modifying objects in a graphical environment
US11733848B2 (en) Emergent content containers
Seligmann Creating a mobile VR interactive tour guide
KR102555968B1 (en) Artificial intelligence-based digital content creation device and method for generating digital content from text information
Xin et al. AR Interaction Design Mode of Multi-user and Multi-character in Theme Parks
Drugge Evaluating the use of mobile Augmented Reality as a digital communication tool
Rodriguez et al. The AR Floating Farm Project Student Guide
Halpine Between Authority and Pretense

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20781224

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20781224

Country of ref document: EP

Kind code of ref document: A1