EP3358462A1 - Assistant de réalité virtuelle (rv) adaptatif basé sur le contexte dans des environnements de rv - Google Patents
Assistant de réalité virtuelle (rv) adaptatif basé sur le contexte dans des environnements de rv Download PDFInfo
- Publication number
- EP3358462A1 EP3358462A1 EP17163301.9A EP17163301A EP3358462A1 EP 3358462 A1 EP3358462 A1 EP 3358462A1 EP 17163301 A EP17163301 A EP 17163301A EP 3358462 A1 EP3358462 A1 EP 3358462A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- input
- environment
- context
- user
- adaptive
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/451—Execution arrangements for user interfaces
- G06F9/453—Help systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/24—Querying
- G06F16/245—Query processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0482—Interaction with lists of selectable items, e.g. menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/451—Execution arrangements for user interfaces
Definitions
- the disclosure herein generally relate to virtual assistant systems, and, more particularly, to context based adaptive virtual reality (VR) assistant for VR environments.
- VR virtual reality
- a method for providing a context based virtual reality assistant in one or more VR environments comprising providing an adaptive virtual reality (VR) assistant application executable by at least one processor configured for VR assistance on a computing device; detecting, by the adaptive VR assistant application, activation of an interface element from a plurality of interface elements; and executing, upon the activation of the interface element, at least one of a first set of instructions and a second set of instructions, wherein the first set of instructions comprises: displaying, in the real time, one or more VR environments upon activation of an interface element; determining, in the real time, a selection of at least one VR environment from the one or more VR environments by a user; displaying, in the real time, the at least one selected VR environment comprising one of one or more corresponding objects and one or more VR characters on the computing device, and generating a VR character for the user
- VR virtual reality
- the first media message and the second media message may comprise at least one of a text message, an audio message, a video message, an audiovisual message, a gaze input, a gesture input, or combination thereof.
- a second interactive communication may be enabled between the user and the corresponding objects in the generated VR environment based on the second determined context.
- text from the first input and the second input are extracted to determine the first context and the second context respectively.
- a computer implemented system for providing a context based virtual reality assistant in one or more VR environments comprising: a memory storing instructions; one or more communication interfaces; and one or more hardware processors coupled to said memory using said one or more communication interfaces, wherein said one or more hardware processors are configured by said instructions to: provide an adaptive virtual reality (VR) assistant application executable by at least one processor configured for VR assistance on a computing device; detect, by the adaptive VR assistant application, activation of an interface element from a plurality of interface elements; and execute, upon the activation of the interface element, at least one of a first set of instructions and a second set of instructions, wherein the first set of instructions comprises: displaying, in the real time, one or more VR environments upon activation of an interface element; determining, in the real time, a selection of at least one VR environment from the one or more VR environments by a user; displaying, in the real time, the at least one selected VR environment comprising one of one or more corresponding objects and one or more VR characters on
- VR virtual reality
- the first media message and the second media message may comprise at least one of a text message, an audio message, a video message, an audiovisual message, a gaze input, a gesture input, or combination thereof.
- the system is further configured to enable a second interactive communication between the user and the corresponding objects in the generated VR environment based on the second determined context.
- text from the first input and the second input are extracted to determine the first context and the second context respectively.
- one or more non-transitory machine readable information storage mediums comprising one or more instructions.
- the one or more instructions which when executed by one or more hardware processors causes providing an adaptive virtual reality (VR) assistant application executable by at least one processor configured for (or for providing) VR assistance on a computing device; detecting, by the adaptive VR assistant application, activation of an interface element from a plurality of interface elements; and executing, upon the activation of the interface element, at least one of a first set of instructions and a second set of instructions, wherein the first set of instructions comprises: displaying, in the real time, one or more VR environments upon activation of an interface element; determining, in the real time, a selection of at least one VR environment from the one or more VR environments by a user; displaying, in the real time, the at least one selected VR environment comprising one of one or more corresponding objects and one or more VR characters on the computing device, and generating a VR character for the user specific to the at least one selected VR environment; obtaining, in the real time
- VR virtual
- the first media message and the second media message may comprise at least one of a text message, an audio message, a video message, an audiovisual message, a gaze input, a gesture input, or combination thereof.
- the instructions further cause the one or more hardware processors to enable a second interactive communication between the user and the corresponding objects in the generated VR environment based on the second determined context.
- text from the first input and the second input are extracted to determine the first context and the second context respectively.
- FIGS. 1 through 4 where similar reference characters denote corresponding features consistently throughout the figures, there are shown preferred embodiments and these embodiments are described in the context of the following exemplary system and/or method.
- FIG. 1 illustrates an exemplary block diagram of a system 100 for providing a virtual reality (VR) and context based virtual assistant in one or more VR environments according to an embodiment of the present disclosure. More specifically, FIG. 1 depicts an exemplary system that provides context based adaptive VR assistant in VR environments.
- the system 100 includes one or more processors 104, communication interface device(s) or input/output (I/O) interface(s) 106, and one or more data storage devices or memory 102 operatively coupled to the one or more processors 104.
- the one or more processors 104 that are hardware processors can be implemented as one or more microprocessors, microcomputers, microcontrollers, digital signal processors, central processing units, state machines, logic circuitries, and/or any devices that manipulate signals based on operational instructions.
- the processor(s) is configured to fetch and execute computer-readable instructions stored in the memory.
- the system 100 can be implemented in a variety of computing systems, such as laptop computers, notebooks, hand-held devices, workstations, mainframe computers, servers, a network cloud and the like.
- the I/O interface device(s) 106 can include a variety of software and hardware interfaces, for example, a web interface, a graphical user interface, and the like and can facilitate multiple communications within a wide variety of networks N/W and protocol types, including wired networks, for example, LAN, cable, etc., and wireless networks, such as WLAN, cellular, or satellite.
- the I/O interface device(s) (106) can include one or more ports for connecting a number of devices to one another or to another server.
- the memory 102 may include any computer-readable medium known in the art including, for example, volatile memory, such as static random access memory (SRAM) and dynamic random access memory (DRAM), and/or non-volatile memory, such as read only memory (ROM), erasable programmable ROM, flash memories, hard disks, optical disks, and magnetic tapes.
- volatile memory such as static random access memory (SRAM) and dynamic random access memory (DRAM)
- non-volatile memory such as read only memory (ROM), erasable programmable ROM, flash memories, hard disks, optical disks, and magnetic tapes.
- ROM read only memory
- erasable programmable ROM erasable programmable ROM
- FIG. 2 illustrates an exemplary flow diagram of a method for providing a virtual reality (VR) and context based virtual assistant in one or more VR environments using the system 100 of FIG. 1 in accordance with an embodiment of the present disclosure.
- the system 100 comprises one or more data storage devices or the memory 102 operatively coupled to the one or more hardware processors 104 and is configured to store instructions for execution of steps of the method by the one or more processors 104.
- the steps of the method of the present disclosure will now be explained with reference to the components of the system 100 as depicted in FIG. 1 , and the flow diagram.
- the one or more processors 104 provide an adaptive virtual reality (VR) assistant application executable by the one or more hardware processors 104 configured for (or for providing) VR assistance on a computing device.
- VR virtual reality
- the expressions 'adaptive VR assistant application', 'VR assistant', 'VR assistant application', and 'VR application' may be interchangeably used herein.
- the one or more hardware processors 104 detect, using the adaptive VR assistant application, activation of an interface element from a plurality of interface elements.
- the plurality of interface elements may include, but not limited to, VR environment interface option, a communication session (or chat) interface option, and the like.
- the one or more hardware processors 104 execute, upon the activation of the interface element, at least one of a first set of instructions and a second set of instructions.
- the first set of instructions comprises displaying (208), in the real time, one or more VR environments upon activation of an interface element.
- the method further includes determining (210), in the real time, a selection of at least one VR environment from the one or more VR environments by a user.
- the one or more hardware processors 104 display, in the real time, the at least one selected VR environment comprising one of one or more corresponding objects and one or more VR characters on the computing device, and generates a VR character for the user specific to the at least one selected VR environment.
- the one or more hardware processors 104 obtain, in the real time, a first input comprising a first media message from the user.
- the one or more hardware processors 104 determine, in the real time, using a Natural Language Processing (NLP) engine, a first context based on the first media message.
- NLP Natural Language Processing
- the Natural Language Processing engine may be stored in the memory 102 and executed on the computing device.
- the one or more hardware processors 104 enable, based on the first determined context, a first interactive communication session between the VR character and one or the corresponding objects and the one or more presented VR characters in the at least one selected VR environment.
- the system 100 may generate a VR assistant that enables a first interactive communication session between the VR character and one or the corresponding objects and the one or more presented VR characters in the at least one selected VR environment.
- the system 100 may directly generate (or enable) a first interactive communication session between the VR character and one or the corresponding objects and the one or more presented VR characters in the at least one selected VR environment, wherein the system 100 may generate a virtual assistant that can help the VR character (e.g., user and other audiences) during interaction.
- the second set of instructions comprises: generating (220) an interactive session user interface (also referred as 'interactive communication session user interface') by the adaptive virtual reality (VR) assistant application on the computing device.
- the second set of instructions further comprises: obtaining (222) a second input comprising one or more queries from the user; determining (224) a second context of the second input; and generating (226), by the adaptive virtual reality (VR) assistant application, at least one of: one or more responses or a VR environment within the generated interactive session user interface, based on one of: the second input, the one or more responses, the determined context, and any combination thereof.
- the system 100 generated one or more responses and/or the VR environment such that the generated responses and/or VR environment are integrated within the generated interactive session user interface.
- the hardware processors may further enable a second interactive communication between the user and the corresponding objects in the generated VR environment based on the second determined context.
- the first media message and the second media message may comprise at least one of a text message, an audio message (e.g., voice), a video message, an audiovisual message, a gaze input, a gesture input, or combination thereof.
- text from the first input and the second input are extracted and processed by the NLP engine to determine the first context and the second context respectively.
- FIG. 3 is an exemplary user interface view of a VR environment generated by the adaptive VR assistant application executed on the system 100 of FIG. 1 according to an embodiment of the present disclosure.
- the adaptive VR assistant application obtains one or more inputs from the user.
- the inputs may comprise, but not limited to media message.
- the adaptive VR assistant application Based on the text extracted from the media message, the adaptive VR assistant application generates (or renders) a VR environment as depicted in FIG. 3 .
- Rendering of a VR environment may comprise selecting and retrieving a determined context based VR environment from one or more VR environments (stored in one or more asset repositories).
- the user may have selected auditorium as a VR environment, and provides attending conference or speaking at an event as input (e.g., the media message) to the adaptive VR assistant application.
- This input is extracted by the adaptive VR assistant application and a context is determined.
- the adaptive VR assistant application generates a VR character for the user (e.g., a VR speaker or a 3D avatar - augmented virtual presence of real world characters), and one or more VR characters (e.g., audience) 304 who would be attending the event (e.g., virtual presence of users), in the selected VR environment 302 (e.g., auditorium).
- the adaptive VR assistant application may generate a VR assistant (e.g., a virtual user) that may analyse the scenario and engage with the user (e.g., VR speaker). Additionally, the adaptive VR assistant application presents, in the real time, the at least one selected VR environment comprising one of one or more corresponding objects and one or more VR characters on the computing device, and generates a VR character for the user specific to the at least one selected VR environment. More specifically, the adaptive VR assistant generates a public speaking simulator that simulates one or more VR environments based on user's selection in which the VR assistant augments the presence of user as a speaker to address the virtual audience. As depicted in FIG. 3 , the user interface depicts a VR environment which involves session with virtual audience and the virtual speaker. The environment which have the virtual audience involves the interaction in which user gives the speech and audience can cross question over it, wherein this cross questioning is the audio that is coming from the NLP engine after context analysis.
- a VR assistant e.g., a virtual user
- the adaptive VR assistant application presents, in
- the adaptive VR assistant application may first receive an input comprising a media message, then extract, using the NLP engine, text from the input and determine the context. Based on the determined context, the adaptive VR assistant application generates a VR environment (e.g., in this scenario: an auditorium) and corresponding objects (e.g., stage/ podium, dais, with microphone, and sound speakers etc.) and VR characters (e.g., audience) and enables an interactive communication session between the virtual audience and the virtual speaker as depicted in FIG. 3 .
- the corresponding objects, virtual character (e.g., virtual speaker), and virtual users (e.g., audience) may be pre-defined and stored in one or more asset repositories (not shown in FIGS.
- the NLP engine (not shown in FIGS. 1-3 ) residing in the memory 102.
- the NLP engine processes the media message (e.g., an audio) to extract text from the audio and then text is processed to understand the context and the final result in the form of audio/text is sent back to an asset loader (or the one or more asset repositories) which load/manipulate the content in a 3D Space (e.g., a 3D virtual reality environment).
- a 3D Space e.g., a 3D virtual reality environment
- the adaptive VR assistant application may further activate one or more functionalities 306 that enable user of the adaptive VR assistant application to perform one or more actions in real time.
- the one or more functionalities may be mapped to one or more VR environments, and the information pertaining to the mapping may be stored in an association database (not shown in FIGS. 1-3 ) residing in the memory 102.
- an association database not shown in FIGS. 1-3
- the user is requested (or when the user focuses) to focus in one direction and for a predefined time interval (period)
- one or more functionalities can be activated (e.g., recording functionality pertaining to an event in the generated VR environment) based on one or more inputs (e.g., but not limited to, gaze inputs).
- the computing device may be paired with a VR gear wherein the VR gear can be worn by the user to experience the VR environment and perform one or more actions.
- FIG. 4 illustrates an exemplary user interface view depicting an interactive session user interface that is integrated with a VR environment using the adaptive VR assistant application of the system 100 of FIG. 1 according to an embodiment of the present disclosure.
- FIG. 4A depicts an interactive session user interface, wherein the user and the VR assistant associated with the adaptive VR assistant application 402 engage in an interactive communication session 404 within a VR environment 406.
- the user provides one or more inputs comprising one or more queries.
- the VR assistant provides one or more responses to the one or more queries obtained from the user, and based on the queries received from the user, determines the context, and generates the VR environment 406. For example, as shown in FIG.
- the user has provided inputs (or queries).
- queries include, but not limited to, 'can you show me a VR view'.
- the VR assistant may either provide responses or ask more questions to user (e.g., 'which class you want to view in VR').
- the user has responded to the query by inputting answer (e.g., business class).
- the adaptive VR assistant has determined the context of the queries and provided (or generated a VR environment that depicts a multi-dimensional (e.g., a 3D VR view) that illustrates an in-flight view of a business class pertaining to the determining context.
- the adaptive VR assistant application generates an in-flight VR environment for a business class within the same interactive communication session of the same user interface as depicted in FIG. 4 .
- the user may provide inputs on the generated VR view mode (e.g., in-flight view) by the adaptive VR assistant application.
- the VR environment e.g., the in-flight view
- the VR view mode may be provided or displayed on a separate user interface (or disintegrate the chat (or communication session interface) and the VR view mode (or generated VR environment that depicts an in-flight view of the business class) thus enabling the user to experience the generated VR environment.
- the adaptive VR assistant application help as an assistant to book a flight, to provide walk-through of basic check-in formalities, booking of lounge etc.
- a specific scene is being loaded by understanding the context of the interactive communication session.
- a user may be asked to attend basic check-in walkthrough in VR thereby providing him/her one or more options to switch into VR mode for the same.
- the embodiments of the present disclosure provide an auditorium scenario, and in-flight scenario in VR environments, it is to be understood to a person skilled (or having ordinary skill) in the art that the embodiments of the present disclosure and the proposed system 100 may be implemented of other scenarios, and should not be construed as limiting the scope of the embodiments herein.
- the embodiments of the present disclosure provide systems and methods that implement adaptive VR assistant application (having a VR assistant) for providing context based assistance in VR environments.
- the system 100 provides an engaging experience to users wherein at both ends the system 100 replicates virtual users.
- the embodiments of the present disclosure provides a way to interact with the real and 3D persons in virtual spaces, by providing manipulation and control over the view of the person (or users) present in 3D Space.
- the proposed system 100 provides a way to make changes in VR Environment and take decisions in real time (e.g., by way of automation and as well manual technique).
- the proposed system 100 understands the context from users' media message and guides them based on the understanding extracted from the context and content. Furthermore, the system 100 enables control and dynamic rendering of the content from a remote user. Moreover, the system 100 may enable users to implement the system 100 in one or more computing devices (e.g., computer systems, laptops, mobile communication devices (by pairing them with one or more VR gears) wherein the users can experience the VR environment. For example, users can practice speaking at events/conferences by implementing the system 100 which generates a virtual reality environment (e.g., auditorium), wherein users can speak and can develop skills (e.g., reducing stage fright, and build confidence).
- a virtual reality environment e.g., auditorium
- the hardware device can be any kind of device which can be programmed including e.g. any kind of computer like a server or a personal computer, or the like, or any combination thereof.
- the device may also include means which could be e.g. hardware means like e.g. an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA), or a combination of hardware and software means, e.g.
- ASIC application-specific integrated circuit
- FPGA field-programmable gate array
- the means can include both hardware means and software means.
- the method embodiments described herein could be implemented in hardware and software.
- the device may also include software means.
- the embodiments may be implemented on different hardware devices, e.g. using a plurality of CPUs.
- the embodiments herein can comprise hardware and software elements.
- the embodiments that are implemented in software include but are not limited to, firmware, resident software, microcode, etc.
- the functions performed by various modules described herein may be implemented in other modules or combinations of other modules.
- a computer-usable or computer readable medium can be any apparatus that can comprise, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
- a computer-readable storage medium refers to any type of physical memory on which information or data readable by a processor may be stored.
- a computer-readable storage medium may store instructions for execution by one or more processors, including instructions for causing the processor(s) to perform steps or stages consistent with the embodiments described herein.
- the term "computer-readable medium” should be understood to include tangible items and exclude carrier waves and transient signals, i.e., be non-transitory. Examples include random access memory (RAM), read-only memory (ROM), volatile memory, nonvolatile memory, hard drives, CD ROMs, DVDs, flash drives, disks, and any other known physical storage media.
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
IN201721004213 | 2017-02-06 |
Publications (1)
Publication Number | Publication Date |
---|---|
EP3358462A1 true EP3358462A1 (fr) | 2018-08-08 |
Family
ID=58454905
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP17163301.9A Pending EP3358462A1 (fr) | 2017-02-06 | 2017-03-28 | Assistant de réalité virtuelle (rv) adaptatif basé sur le contexte dans des environnements de rv |
Country Status (2)
Country | Link |
---|---|
US (1) | US11113080B2 (fr) |
EP (1) | EP3358462A1 (fr) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109324687A (zh) * | 2018-08-14 | 2019-02-12 | 华为技术有限公司 | 一种显示方法以及虚拟现实设备 |
CN109584376A (zh) * | 2018-12-03 | 2019-04-05 | 广东工业大学 | 基于vr技术的构图教学方法、装置、设备以及存储介质 |
CN114356083A (zh) * | 2021-12-22 | 2022-04-15 | 阿波罗智联(北京)科技有限公司 | 虚拟个人助理控制方法、装置、电子设备及可读存储介质 |
Families Citing this family (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8676904B2 (en) | 2008-10-02 | 2014-03-18 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
CN113470640B (zh) | 2013-02-07 | 2022-04-26 | 苹果公司 | 数字助理的语音触发器 |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
DK180048B1 (en) | 2017-05-11 | 2020-02-04 | Apple Inc. | MAINTAINING THE DATA PROTECTION OF PERSONAL INFORMATION |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
US20180331841A1 (en) * | 2017-05-12 | 2018-11-15 | Tsunami VR, Inc. | Systems and methods for bandwidth optimization during multi-user meetings that use virtual environments |
DK201770429A1 (en) | 2017-05-12 | 2018-12-14 | Apple Inc. | LOW-LATENCY INTELLIGENT AUTOMATED ASSISTANT |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
DK180639B1 (en) | 2018-06-01 | 2021-11-04 | Apple Inc | DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
DK201970509A1 (en) | 2019-05-06 | 2021-01-15 | Apple Inc | Spoken notifications |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
US11468890B2 (en) | 2019-06-01 | 2022-10-11 | Apple Inc. | Methods and user interfaces for voice-based control of electronic devices |
US11372474B2 (en) | 2019-07-03 | 2022-06-28 | Saec/Kinetic Vision, Inc. | Systems and methods for virtual artificial intelligence development and testing |
EP4075411A4 (fr) * | 2020-04-24 | 2023-01-18 | Samsung Electronics Co., Ltd. | Dispositif et procédé de fourniture de simulation de public interactive |
US11010129B1 (en) | 2020-05-08 | 2021-05-18 | International Business Machines Corporation | Augmented reality user interface |
US11061543B1 (en) | 2020-05-11 | 2021-07-13 | Apple Inc. | Providing relevant data items based on context |
US11490204B2 (en) | 2020-07-20 | 2022-11-01 | Apple Inc. | Multi-device audio adjustment coordination |
US11438683B2 (en) | 2020-07-21 | 2022-09-06 | Apple Inc. | User identification using headphones |
US11438551B2 (en) * | 2020-09-15 | 2022-09-06 | At&T Intellectual Property I, L.P. | Virtual audience using low bitrate avatars and laughter detection |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6091410A (en) * | 1997-11-26 | 2000-07-18 | International Business Machines Corporation | Avatar pointing mode |
US20140317502A1 (en) * | 2013-04-18 | 2014-10-23 | Next It Corporation | Virtual assistant focused user interfaces |
US20150089373A1 (en) * | 2013-09-20 | 2015-03-26 | Oracle International Corporation | Context aware voice interface for computing devices |
WO2017044257A1 (fr) * | 2015-09-08 | 2017-03-16 | Apple Inc. | Assistant automatisé intelligent dans un environnement multimédia |
WO2017122900A1 (fr) * | 2016-01-14 | 2017-07-20 | Samsung Electronics Co., Ltd. | Appareil et procédé d'utilisation d'agent personnel |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20010055039A1 (en) * | 1996-09-30 | 2001-12-27 | Koichi Matsuda | Three-dimensional virtual reality space display processing apparatus, a three-dimensional virtual reality space display processing method, and an information providing medium |
WO2002057896A2 (fr) | 2001-01-22 | 2002-07-25 | Digital Animations Group Plc | Assistant virtuel interactif |
US20060074831A1 (en) | 2004-09-20 | 2006-04-06 | Hyder Andrew D | Virtual assistant |
US8156060B2 (en) | 2008-02-27 | 2012-04-10 | Inteliwise Sp Z.O.O. | Systems and methods for generating and implementing an interactive man-machine web interface based on natural language processing and avatar virtual agent based character |
US20150074524A1 (en) | 2013-09-10 | 2015-03-12 | Lenovo (Singapore) Pte. Ltd. | Management of virtual assistant action items |
US10446055B2 (en) * | 2014-08-13 | 2019-10-15 | Pitchvantage Llc | Public speaking trainer with 3-D simulation and real-time feedback |
US10120413B2 (en) * | 2014-09-11 | 2018-11-06 | Interaxon Inc. | System and method for enhanced training using a virtual reality environment and bio-signal data |
US20180101986A1 (en) * | 2016-10-10 | 2018-04-12 | Aaron Mackay Burns | Drawing in a 3d virtual reality environment |
US10861344B2 (en) * | 2017-01-31 | 2020-12-08 | Cerego, Llc. | Personalized learning system and method for the automated generation of structured learning assets based on user data |
US20180218374A1 (en) * | 2017-01-31 | 2018-08-02 | Moveworks, Inc. | Method, system and computer program product for facilitating query resolutions at a service desk |
-
2017
- 2017-03-28 EP EP17163301.9A patent/EP3358462A1/fr active Pending
- 2017-03-29 US US15/473,034 patent/US11113080B2/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6091410A (en) * | 1997-11-26 | 2000-07-18 | International Business Machines Corporation | Avatar pointing mode |
US20140317502A1 (en) * | 2013-04-18 | 2014-10-23 | Next It Corporation | Virtual assistant focused user interfaces |
US20150089373A1 (en) * | 2013-09-20 | 2015-03-26 | Oracle International Corporation | Context aware voice interface for computing devices |
WO2017044257A1 (fr) * | 2015-09-08 | 2017-03-16 | Apple Inc. | Assistant automatisé intelligent dans un environnement multimédia |
WO2017122900A1 (fr) * | 2016-01-14 | 2017-07-20 | Samsung Electronics Co., Ltd. | Appareil et procédé d'utilisation d'agent personnel |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109324687A (zh) * | 2018-08-14 | 2019-02-12 | 华为技术有限公司 | 一种显示方法以及虚拟现实设备 |
CN109324687B (zh) * | 2018-08-14 | 2021-10-01 | 华为技术有限公司 | 一种显示方法以及虚拟现实设备 |
US11748950B2 (en) | 2018-08-14 | 2023-09-05 | Huawei Technologies Co., Ltd. | Display method and virtual reality device |
CN109584376A (zh) * | 2018-12-03 | 2019-04-05 | 广东工业大学 | 基于vr技术的构图教学方法、装置、设备以及存储介质 |
CN114356083A (zh) * | 2021-12-22 | 2022-04-15 | 阿波罗智联(北京)科技有限公司 | 虚拟个人助理控制方法、装置、电子设备及可读存储介质 |
Also Published As
Publication number | Publication date |
---|---|
US11113080B2 (en) | 2021-09-07 |
US20180225131A1 (en) | 2018-08-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11113080B2 (en) | Context based adaptive virtual reality (VR) assistant in VR environments | |
US20180122368A1 (en) | Multiparty conversation assistance in mobile devices | |
CN112313921A (zh) | 促进会议加入 | |
US10187439B2 (en) | Dynamic recording of online conference | |
US10971168B2 (en) | Dynamic communication session filtering | |
US20220138470A1 (en) | Techniques for Presentation Analysis Based on Audience Feedback, Reactions, and Gestures | |
US10699709B2 (en) | Conference call analysis and automated information exchange | |
US20220141532A1 (en) | Techniques for rich interaction in remote live presentation and accurate suggestion for rehearsal through audience video analysis | |
CN111556279A (zh) | 即时会话的监控方法和通信方法 | |
CN109948151A (zh) | 构建语音助手的方法 | |
US11792468B1 (en) | Sign language interpreter view within a communication session | |
WO2022125341A1 (fr) | Promotion d'utilisateurs dans des sessions de collaboration | |
US11195619B2 (en) | Real time sensor attribute detection and analysis | |
US11716364B2 (en) | Reducing bandwidth requirements of virtual collaboration sessions | |
US11575528B1 (en) | Dynamic virtual meeting, online webinar and distance-based education systems and methods | |
US11769504B2 (en) | Virtual meeting content enhancement triggered by audio tracking | |
US20230061210A1 (en) | Method and system of automated question generation for speech assistance | |
CN113938697A (zh) | 直播间内的虚拟发言方法、装置及计算机设备 | |
US11954778B2 (en) | Avatar rendering of presentations | |
US11729009B1 (en) | Intelligent meeting agent | |
US11888907B2 (en) | Workflow-based screen sharing during live presentation | |
US11526669B1 (en) | Keyword analysis in live group breakout sessions | |
US20230245658A1 (en) | Asynchronous pipeline for artificial intelligence service requests | |
US20230327895A1 (en) | Content viewing guidance in an online meeting | |
US20220391044A1 (en) | Dynamic presentation adjustment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION HAS BEEN PUBLISHED |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
AX | Request for extension of the european patent |
Extension state: BA ME |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20190208 |
|
RBV | Designated contracting states (corrected) |
Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
17Q | First examination report despatched |
Effective date: 20190730 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |