CN117632391A - Application control method, device, equipment and medium based on virtual reality space - Google Patents

Application control method, device, equipment and medium based on virtual reality space Download PDF

Info

Publication number
CN117632391A
CN117632391A CN202210964215.XA CN202210964215A CN117632391A CN 117632391 A CN117632391 A CN 117632391A CN 202210964215 A CN202210964215 A CN 202210964215A CN 117632391 A CN117632391 A CN 117632391A
Authority
CN
China
Prior art keywords
application
virtual reality
preset
reality space
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210964215.XA
Other languages
Chinese (zh)
Inventor
杨飘
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Zitiao Network Technology Co Ltd
Original Assignee
Beijing Zitiao Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Zitiao Network Technology Co Ltd filed Critical Beijing Zitiao Network Technology Co Ltd
Priority to CN202210964215.XA priority Critical patent/CN117632391A/en
Publication of CN117632391A publication Critical patent/CN117632391A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/48Program initiating; Program switching, e.g. by interrupt
    • G06F9/4806Task transfer initiation or dispatching
    • G06F9/4843Task transfer initiation or dispatching by program, e.g. task dispatcher, supervisor, operating system
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The embodiment of the disclosure relates to an application control method, device, equipment and medium based on virtual reality space, wherein the method comprises the following steps: responding to temporary exit operation of the target application, closing an application interface of the target application in the virtual reality space, and switching an application process of the target application into background process operation; displaying an application identification model corresponding to the target application in the virtual reality space; and responding to a first preset triggering operation of the application identification model, switching the application process into a foreground process operation, and displaying an application interface of the target application in the virtual reality space. In the embodiment provided by the disclosure, application suspension in the virtual reality space is realized, the path of application entering is shortened, and the sense of reality of application control is improved.

Description

Application control method, device, equipment and medium based on virtual reality space
Technical Field
The disclosure relates to the technical field of virtual reality, and in particular relates to an application control method, device, equipment and medium based on virtual reality space.
Background
Virtual Reality (VR) technology, also known as Virtual environments, moods, or artificial environments, refers to technology that utilizes a computer to generate a Virtual world that can directly impart visual, auditory, and tactile sensations to participants and allow them to interactively observe and operate. The improvement of VR realism to make the experience of virtual reality space and real physical space feel similar becomes a mainstream.
In the related technology, the operation of the application can be realized based on the virtual reality technology, the application interface corresponding to the application is rendered in the virtual reality space, the suspension operation of the application is realized, and the like.
However, in the virtual reality technology, only the process of the corresponding application can be directly closed, and after the next time an opening instruction of the application is detected, the process of the application needs to be restarted to realize the re-opening of the application, so that the application has a longer access path.
Disclosure of Invention
In order to solve the above technical problems or at least partially solve the above technical problems, the present disclosure provides an application control method, device, equipment and medium based on a virtual reality space, so as to solve the technical problem in the prior art that in the virtual reality space, only a process of a corresponding application can be directly closed, and when an opening instruction of the application is detected next time, the process of the application needs to be pulled up again to realize reopening of the application, which results in a longer application access path.
The embodiment of the disclosure provides an application control method based on a virtual reality space, which comprises the following steps: responding to temporary exit operation of a target application, and switching an application process of the target application into background process operation, wherein when the application process of the target application is switched into background process operation, an application interface of the target application is stopped from being displayed in the virtual reality space; displaying an application identification model corresponding to the target application in the virtual reality space; and switching the application process into a foreground process operation in response to a first preset trigger operation on the application identification model, wherein when the application process is switched into the foreground process operation, an application interface of the target application is displayed in the virtual reality space.
The embodiment of the disclosure also provides an application control device based on the virtual reality space, which comprises: the process control module is used for responding to the temporary exit operation of the target application and switching the application process of the target application into background process operation, wherein when the application process of the target application is switched into background process operation, the application interface of the target application is stopped from being displayed in the virtual reality space; the display module is used for displaying an application identification model corresponding to the target application in the virtual reality space; the process control module is further configured to switch the application process to a foreground process operation in response to a first preset trigger operation on the application identification model, where when the application process is switched to the foreground process operation, an application interface of the target application is displayed in the virtual reality space.
The embodiment of the disclosure also provides an electronic device, which comprises: a processor; a memory for storing the processor-executable instructions; the processor is configured to read the executable instructions from the memory and execute the instructions to implement the application control method based on the virtual reality space according to the embodiment of the disclosure.
The present disclosure also provides a computer-readable storage medium storing a computer program for executing the virtual reality space-based application control method as provided by the embodiments of the present disclosure.
Compared with the prior art, the technical scheme provided by the embodiment of the disclosure has the following advantages:
according to the application control scheme based on the virtual reality space, the application process of the target application is switched to the background process operation in response to the temporary exit operation of the target application, wherein when the application process of the target application is switched to the background process operation, the application interface of the target application is stopped to be displayed in the virtual reality space, the application identification model corresponding to the target application is displayed in the virtual reality space, and further, the application process is switched to the foreground process operation in response to the first preset trigger operation of the application identification model, and when the application process is switched to the foreground process operation, the application interface of the target application is displayed in the virtual reality space. Therefore, application suspension in the virtual reality space is realized, the path of application entering is shortened, and the sense of reality of application control is improved.
Drawings
The above and other features, advantages, and aspects of embodiments of the present disclosure will become more apparent by reference to the following detailed description when taken in conjunction with the accompanying drawings. The same or similar reference numbers will be used throughout the drawings to refer to the same or like elements. It should be understood that the figures are schematic and that elements and components are not necessarily drawn to scale.
Fig. 1 is a schematic view of an application scenario of a virtual reality device according to an embodiment of the present disclosure;
fig. 2 is a schematic view of an operation panel display in a virtual reality space according to an embodiment of the disclosure;
fig. 3 is a schematic flow chart of an application control method based on a virtual reality space according to an embodiment of the disclosure;
FIG. 4 is a schematic diagram of an application identifier model according to an embodiment of the disclosure;
fig. 5 is a schematic view of a display scenario of an application identifier model according to an embodiment of the present disclosure;
FIG. 6 is a schematic diagram of a switching animation display scenario provided in an embodiment of the present disclosure;
fig. 7 is a schematic diagram of an application control scenario based on a virtual reality space according to an embodiment of the present disclosure;
fig. 8 is a schematic structural diagram of an application control device based on a virtual reality space according to an embodiment of the disclosure;
fig. 9 is a schematic structural diagram of an electronic device according to an embodiment of the disclosure.
Detailed Description
Embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While certain embodiments of the present disclosure have been shown in the accompanying drawings, it is to be understood that the present disclosure may be embodied in various forms and should not be construed as limited to the embodiments set forth herein, but are provided to provide a more thorough and complete understanding of the present disclosure. It should be understood that the drawings and embodiments of the present disclosure are for illustration purposes only and are not intended to limit the scope of the present disclosure.
It should be understood that the various steps recited in the method embodiments of the present disclosure may be performed in a different order and/or performed in parallel. Furthermore, method embodiments may include additional steps and/or omit performing the illustrated steps. The scope of the present disclosure is not limited in this respect.
The term "including" and variations thereof as used herein are intended to be open-ended, i.e., including, but not limited to. The term "based on" is based at least in part on. The term "one embodiment" means "at least one embodiment"; the term "another embodiment" means "at least one additional embodiment"; the term "some embodiments" means "at least some embodiments. Related definitions of other terms will be given in the description below.
It should be noted that the terms "first," "second," and the like in this disclosure are merely used to distinguish between different devices, modules, or units and are not used to define an order or interdependence of functions performed by the devices, modules, or units.
It should be noted that references to "one", "a plurality" and "a plurality" in this disclosure are intended to be illustrative rather than limiting, and those of ordinary skill in the art will appreciate that "one or more" is intended to be understood as "one or more" unless the context clearly indicates otherwise.
The names of messages or information interacted between the various devices in the embodiments of the present disclosure are for illustrative purposes only and are not intended to limit the scope of such messages or information.
Some technical concepts or noun concepts referred to herein are described in association with:
the virtual reality device, the terminal for realizing the virtual reality effect, may be provided in the form of glasses, a head mounted display (Head Mount Display, HMD), or a contact lens for realizing visual perception and other forms of perception, but the form of the virtual reality device is not limited to this, and may be further miniaturized or enlarged as needed.
The virtual reality device described in the embodiments of the present invention may include, but is not limited to, the following types:
a computer-side virtual reality (PCVR) device performs related computation of a virtual reality function and data output by using a PC side, and an external computer-side virtual reality device realizes a virtual reality effect by using data output by the PC side.
The mobile virtual reality device supports setting up a mobile terminal (such as a smart phone) in various manners (such as a head-mounted display provided with a special card slot), performing related calculation of a virtual reality function by the mobile terminal through connection with the mobile terminal in a wired or wireless manner, and outputting data to the mobile virtual reality device, for example, watching a virtual reality video through an APP of the mobile terminal.
The integrated virtual reality device has a processor for performing the calculation related to the virtual function, and thus has independent virtual reality input and output functions, and is free from connection with a PC or a mobile terminal, and has high degree of freedom in use.
Virtual reality objects, objects that interact in a virtual scene, objects that are stationary, moving, and performing various actions in a virtual scene, such as virtual persons corresponding to a user in a live scene, are controlled by a user or a robot program (e.g., an artificial intelligence based robot program).
As shown in fig. 1, HMDs are relatively light, ergonomically comfortable, and provide high resolution content with low latency. The sensor (such as a nine-axis sensor) for detecting the gesture in the virtual reality device is arranged in the virtual reality device, and is used for detecting the gesture change of the virtual reality device in real time, if the user wears the virtual reality device, when the gesture of the head of the user changes, the real-time gesture of the head is transmitted to the processor, so that the gaze point of the sight of the user in the virtual environment is calculated, an image in the gaze range (namely a virtual view field) of the user in the three-dimensional model of the virtual environment is calculated according to the gaze point, and the image is displayed on the display screen, so that the user looks like watching in the real environment.
In this embodiment, when a user wears the HMD device and opens a predetermined application program, for example, a live video application program, the HMD device may run corresponding virtual scenes, where the virtual scenes may be simulation environments for the real world, semi-simulation virtual scenes, or pure virtual scenes. The virtual scene may be any one of a two-dimensional virtual scene, a 2.5-dimensional virtual scene or a three-dimensional virtual scene, and the dimension of the virtual scene is not limited in the embodiment of the present application. For example, the virtual scene may include characters, sky, land, sea, etc., the land may include environmental elements such as desert, city, etc., the user may control the virtual object to move in the virtual scene, and may also interactively control the controls, models, presentations, characters, etc. in the virtual scene by means of a handle device, a bare hand gesture, etc.
In order to solve the above problems, the present disclosure proposes a technical solution that an application can be suspended in a virtual reality space, so as to simulate temporary exit of the application in reality, and to operate the application in the virtual reality space in a background by adopting a temporary exit mode, and when needed, the application can be directly pulled to be displayed on a foreground without re-entering the corresponding application, thereby improving the entry efficiency of the application and the realism of application control.
The embodiment of the disclosure provides an application control method based on a virtual reality space, in which an operation panel is displayed on a display screen in the virtual reality space, and the operation panel may include entries of various application programs, where the application program entries may be in the form of application program icons (application program icon 1, application program icon 2, application program icon 3, application program icon 4, etc. in the drawing) as shown in fig. 2, etc., and if the corresponding application program entry (application program icon of the application program 1 in the drawing is triggered), the application program entry is triggered, and the application program entry is communicated with a server of the corresponding application program to implement foreground rendering of the application program data on the display screen in the virtual reality space, so as to implement foreground display effect of the application program, etc. That is, in this embodiment, control operations of applications in a virtual reality space can be implemented, which simulates control operations of applications on a display screen such as a mobile phone in a display space.
The method is described below in connection with specific examples.
Fig. 3 is a flow chart of an application control method based on a virtual reality space according to an embodiment of the disclosure, where the method may be performed by an application control device based on a virtual reality space, and the device may be implemented by software and/or hardware, and may be generally integrated in an electronic device. As shown in fig. 3, the method includes:
step 301, switching an application process of the target application to a background process operation in response to the temporary exit operation of the target application, wherein when the application process of the target application is switched to the background process operation, displaying an application interface of the target application is stopped in the virtual reality space.
The target application may be understood as an application corresponding to an application interface currently displayed in the virtual reality space.
In different application scenarios, the manner of responding to the temporary exit operation of the target application is different, and examples are as follows:
in some possible embodiments, the temporary exit operation corresponds to temporarily suspending the target application, similar to switching the corresponding application to a background run or the like while the handset is operating. For example, the user may trigger and input a temporary exit operation through a preset button on the control device (such as a handle device, etc.), detect a trigger operation on a preset operation control (where the preset operation control may be any preset control on the control device) on the control device, and respond to the trigger time period of the trigger operation longer than a preset duration, that is, if a long-press operation on the preset operation control is detected, detect a temporary exit operation on the target application.
In some possible embodiments, there may be other various alternative ways to perform the temporary exit operation, and compared with the way of using the entity device button to perform the temporary exit operation, the alternative way proposes that the operation is not performed by means of the entity device button, which can improve the technical problem that the entity device button is easy to be damaged, and thus the user operation is easy to be affected.
In the optional mode, image information shot by a camera on a user can be monitored to obtain user gesture information, and if the user gesture information is matched with preset gesture information, temporary exit operation on a target application is obtained, wherein the preset gesture information is preset gesture information for temporarily exiting the application.
Further, in response to the temporary exit operation, switching the application process of the target application to a background process operation, wherein when the application process of the target application is switched to the background process operation, similar to the operation in the real space, an application interface of the target application in the virtual reality space is closed.
It should be understood that when the application process of the target application is switched to the background process, the application interface of the target application is hidden or closed, i.e. the application interface is stopped to be displayed in the foreground, but the application process is switched to the background process instead of being exited, and at this time, the communication link with the server of the target application is still maintained, so that the temporary exit of the target application is realized.
Step 302, displaying an application identification model corresponding to the target application in the virtual reality space.
In some possible embodiments, as shown in fig. 4, the application identification model may be a "sphere" model component, and if the target application running in the background is multiple, identification information (not shown in the drawing) of the corresponding target application may be displayed in the application identification model, where the identification information of the target application may be an application icon, an application name, or the like, so that the application identification model may be selectively triggered, and pull-up of the specific target application may be implemented.
In one embodiment of the present disclosure, in order to facilitate intuitive learning that a target application is in a temporary exit state and facilitate subsequent pulling of the temporarily exited target application, an application identification model corresponding to the target application is displayed in a virtual reality space.
The display position of the application identification model may be on the operation panel in the virtual reality space, or may be other positions than the operation panel for improving the intelligent sense of display.
In some possible embodiments, the panel position where the operation panel is located in the virtual reality space is obtained, and the target display position corresponding to the preset panel position is determined, that is, the target display position is placed outside the operation panel, so that the "suspended" sense of display is improved, the intelligent sense of display is improved, and the preset panel position corresponds to the viewpoint of the user, so that the target display position corresponds to the preset panel position, the target display position in the virtual reality space is ensured, and when the application identification model is displayed, the application identification model can be noticed by the user, and the convenience of operation is improved.
It should be noted that, in different application scenarios, the manner of determining the target display position corresponding to the preset panel position is different, and the following is illustrated as an example:
in some possible embodiments, an associated area of a preset azimuth located at a preset panel position is determined, where the preset azimuth may be any direction of the preset panel position, the associated area is a preset area under the preset azimuth, in order to ensure convenience of operation, the distance between the associated area and the operation panel is generally smaller, after the associated area is determined, a target display position is determined in the associated area, for example, a preset display position of the associated area is determined as a target display position, for example, according to a display size of an application identification model, any blank area of the associated area is determined as a target display area.
The display size of the application identification model is generally determined by the distance between the user and the display screen in the virtual reality space, wherein, in order to improve the sense of reality of display, the larger the distance between the user and the display screen in the virtual reality space is, the larger the display size of the corresponding application identification model is, so as to realize the visual display effect of 'near-large-far-small', and of course, the display size of the application identification model can also be calibrated in advance.
In this embodiment, as shown in fig. 5, if the preset direction is the right direction of the operation panel, the application identification model may be displayed in the right area of the operation panel, and the application identification model displayed in the right direction is not only in line with the operation habit of the user, but also convenient for the user to notice.
In some possible embodiments, the target display position may be determined according to the preset azimuth and the preset distance, where the preset azimuth and the preset distance of the preset panel position correspond to each other.
In other possible embodiments, the user may further specify the target display position in the virtual reality space through the manipulation device according to personal preference, for example, determine the current manipulation position of the manipulation device on the display screen in the virtual reality space, where the manipulation position may be implemented through operations such as rotation of the manipulation device, and in response to the user triggering the preset selection control on the manipulation device at the corresponding manipulation position, the determined current manipulation position is the target display position. In order to more intuitively indicate the manipulation position, with continued reference to fig. 5, the manipulation position of the manipulation device on the display screen in the virtual reality space currently may be indicated in a "ray" model manner.
And step 303, switching the application process to a foreground process operation in response to a first preset trigger operation of the application identification model, wherein when the application process is switched to the foreground process operation, an application interface of the target application is displayed in the virtual reality space.
In one embodiment of the present disclosure, a process running in the background may be quickly entered without re-opening the target application, and in this embodiment, in response to a first preset trigger operation of the application identification model, the first trigger operation may be understood as a pull-up operation on the target application.
In this embodiment, in response to a first preset trigger operation of the application identifier model, the application process is switched to a foreground process for running, where when the application process is switched to the foreground process for running, an application interface of the target application is displayed in the virtual reality space, and when the application interface is switched to the background for running, the last displayed application interface of the target application is to be continued to run in the foreground according to the state when the target application is suspended, where the switching of the application process to the foreground process for running is similar to the switching of the application process in the display world from the background to the foreground for running, and in this embodiment, the current application information corresponding to the application process is rendered to the display screen in the virtual reality space for continued display.
In some possible embodiments, in order to enhance the intelligence of the display, when the application interface of the target application is displayed in the virtual reality space, a preset switching animation may be displayed in the virtual reality space, and after the preset switching animation is displayed, the application interface may be displayed.
The preset switching animation may be any preset animation that shows a switching transition effect, and in some optional embodiments, in order to show the relevance between the application identification model and the application interface of the target application, the preset switching animation may be displayed on the application identification model, for example, the preset switching animation is displayed by changing the color of the application identification model according to a preset display period; or, a mapping animation of the application identification model may be displayed, where the mapping animation may be in an animation form that visually associates the application identification model with the application interface, such as "arrow launch animation" shown in fig. 6, so that the application interface is displayed after the preset switching animation display is finished, thereby implementing switching of the target application to the foreground operation.
In one embodiment of the present disclosure, an application interface of a target application may be displayed on an operation panel, or may be displayed at any position in a virtual reality space in a floating window as shown in fig. 7, where a display layer of the application interface is different from a layer of the operation panel, so as to form an intelligent sense of "floating" display.
In an embodiment of the present disclosure, to further realize a sense of realism of an application operation, a closing operation of an application in the real world may be further simulated, where a click mark space position of a preset control device is identified, where the click mark space position corresponding to the control device may be indicated in a virtual reality space by an end point of a model such as a "ray", and in response to the click mark space position matching a space position of a closing control of an application interface of a target application (for example, the closing control shown in fig. 7), an application process of the target application is closed, after the application process of the target application, the application interface of the target application is closed, and at this time, the application process is completely cleared, and a memory occupied by the application process is released.
Of course, in the actual execution process, the user may want to directly terminate the suspension process running in the background, and may directly close the corresponding application process through the second preset trigger operation of the application identification model on the premise of no longer entering the corresponding application.
In this embodiment, an application process of the target application is closed in response to a second preset trigger operation of the application identification model, where after the application process of the target application is closed, an application interface of the target application is closed.
The method comprises the steps of switching an application process from a background to a foreground due to a first preset triggering operation of an application identification model, and closing the corresponding application process due to a second preset triggering operation of the application identification model. Therefore, the first preset trigger operation and the second preset trigger operation are different, and the trigger types of the first preset trigger operation and the second preset trigger operation can be similar, so that the learning cost of a user for learning the trigger operations of different types is reduced.
In one embodiment of the present disclosure, a click flag spatial location of a preset manipulation device is identified, and in the event that the click flag spatial location matches a spatial location of an application identification model, a click operation of a preset operation control on the manipulation device is responded.
In this embodiment, a click flag spatial position of a preset control device is identified, and in a case where the click flag spatial position matches a spatial position of an application flag model, a long-press operation on a preset operation control is responded.
In one embodiment of the present disclosure, the "close" and "open" controls may also be displayed in the associated area of the virtual reality space by applying the identification model, and by triggering the "close" or "open" controls, closing the application process is achieved, or the application process is switched to the foreground display.
In summary, according to the application control method based on the virtual reality space in the embodiment of the disclosure, in response to a temporary exit operation on a target application, an application process of the target application is switched to a background process operation, wherein when the application process of the target application is switched to the background process operation, an application interface of the target application is stopped to be displayed in the virtual reality space, an application identification model corresponding to the target application is displayed in the virtual reality space, and in turn, in response to a first preset trigger operation on the application identification model, the application process is switched to a foreground process operation, and when the application process is switched to the foreground process operation, the application interface of the target application is displayed in the virtual reality space. Therefore, application suspension in the virtual reality space is realized, the path of application entering is shortened, and the sense of reality of application control is improved.
In order to achieve the above embodiments, the present disclosure further provides an application control device based on a virtual reality space.
Fig. 8 is a schematic structural diagram of an application control device based on a virtual reality space according to an embodiment of the disclosure, where the device may be implemented by software and/or hardware, and may be generally integrated in an electronic device to implement application control based on the virtual reality space. As shown in fig. 8, the apparatus includes: a process control module 810, and a display module 820, wherein,
in one embodiment of the present disclosure, the process control module 810 switches an application process of a target application to a background process operation in response to a temporary exit operation for the target application, where when the application process of the target application is switched to the background process operation, displaying an application interface of the target application is stopped in a virtual reality space;
in one embodiment of the present disclosure, the display module 820 displays an application identification model corresponding to a target application in a virtual reality space;
in one embodiment of the present disclosure, the process control module 810 is further configured to switch the application process to a foreground process operation in response to a first preset trigger operation of the application identification model, where when the application process is switched to the foreground process operation, an application interface of the target application is displayed in the virtual reality space.
The application control device based on the virtual reality space provided by the embodiment of the disclosure can execute the application control method based on the virtual reality space provided by any embodiment of the disclosure, and has the corresponding functional modules and beneficial effects of the execution method.
To achieve the above embodiments, the present disclosure also proposes a computer program product comprising a computer program/instruction which, when executed by a processor, implements the virtual reality space based application control method in the above embodiments.
Fig. 9 is a schematic structural diagram of an electronic device according to an embodiment of the disclosure.
Referring now in particular to fig. 9, a schematic diagram of an electronic device 900 suitable for use in implementing embodiments of the present disclosure is shown. The electronic device 900 in the embodiments of the present disclosure may include, but is not limited to, mobile terminals such as mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablet computers), PMPs (portable multimedia players), in-vehicle terminals (e.g., in-vehicle navigation terminals), and the like, and stationary terminals such as digital TVs, desktop computers, and the like. The electronic device shown in fig. 9 is merely an example, and should not impose any limitations on the functionality and scope of use of embodiments of the present disclosure.
As shown in fig. 9, the electronic device 900 may include a processor (e.g., a central processing unit, a graphics processor, etc.) 901, which may perform various appropriate actions and processes in accordance with programs stored in a Read Only Memory (ROM) 902 or programs loaded from a memory 908 into a Random Access Memory (RAM) 903. In the RAM 903, various programs and data necessary for the operation of the electronic device 900 are also stored. The processor 901, the ROM 902, and the RAM 903 are connected to each other by a bus 904. An input/output (I/O) interface 905 is also connected to the bus 904.
In general, the following devices may be connected to the I/O interface 905: input devices 906 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, and the like; an output device 907 including, for example, a Liquid Crystal Display (LCD), a speaker, a vibrator, and the like; memory 908 including, for example, magnetic tape, hard disk, etc.; and a communication device 909. The communication means 909 may allow the electronic device 900 to communicate wirelessly or by wire with other devices to exchange data. While fig. 9 shows an electronic device 900 having various means, it is to be understood that not all illustrated means are required to be implemented or provided. More or fewer devices may be implemented or provided instead.
In particular, according to embodiments of the present disclosure, the processes described above with reference to flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a non-transitory computer readable medium, the computer program comprising program code for performing the method shown in the flow chart. In such an embodiment, the computer program may be downloaded and installed from the network via the communication means 909, or installed from the memory 908, or installed from the ROM 902. When the computer program is executed by the processor 901, the above-described functions defined in the virtual reality space based application control method of the embodiment of the present disclosure are performed.
It should be noted that the computer readable medium described in the present disclosure may be a computer readable signal medium or a computer readable storage medium, or any combination of the two. The computer readable storage medium can be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the foregoing. More specific examples of the computer-readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this disclosure, a computer-readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In the present disclosure, however, the computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave, with the computer-readable program code embodied therein. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination of the foregoing. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: electrical wires, fiber optic cables, RF (radio frequency), and the like, or any suitable combination of the foregoing.
In some implementations, the clients, servers may communicate using any currently known or future developed network protocol, such as HTTP (HyperText Transfer Protocol ), and may be interconnected with any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include a local area network ("LAN"), a wide area network ("WAN"), the internet (e.g., the internet), and peer-to-peer networks (e.g., ad hoc peer-to-peer networks), as well as any currently known or future developed networks.
The computer readable medium may be contained in the electronic device; or may exist alone without being incorporated into the electronic device.
The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to: and switching the application process of the target application into a background process operation in response to the temporary exit operation of the target application, wherein when the application process of the target application is switched into the background process operation, the application interface of the target application is stopped to be displayed in the virtual reality space, an application identification model corresponding to the target application is displayed in the virtual reality space, and further, in response to the first preset trigger operation of the application identification model, the application process is switched into a foreground process operation, and when the application process is switched into the foreground process operation, the application interface of the target application is displayed in the virtual reality space. Therefore, application suspension in the virtual reality space is realized, the path of application entering is shortened, and the sense of reality of application control is improved.
The electronic device may write computer program code for performing the operations of the present disclosure in one or more programming languages, including, but not limited to, an object oriented programming language such as Java, smalltalk, C ++ and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computer (for example, through the Internet using an Internet service provider).
The flowcharts and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units involved in the embodiments of the present disclosure may be implemented by means of software, or may be implemented by means of hardware. Wherein the names of the units do not constitute a limitation of the units themselves in some cases.
The functions described above herein may be performed, at least in part, by one or more hardware logic components. For example, without limitation, exemplary types of hardware logic components that may be used include: a Field Programmable Gate Array (FPGA), an Application Specific Integrated Circuit (ASIC), an Application Specific Standard Product (ASSP), a system on a chip (SOC), a Complex Programmable Logic Device (CPLD), and the like.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. The machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
The foregoing description is only of the preferred embodiments of the present disclosure and description of the principles of the technology being employed. It will be appreciated by persons skilled in the art that the scope of the disclosure referred to in this disclosure is not limited to the specific combinations of features described above, but also covers other embodiments which may be formed by any combination of features described above or equivalents thereof without departing from the spirit of the disclosure. Such as those described above, are mutually substituted with the technical features having similar functions disclosed in the present disclosure (but not limited thereto).
Moreover, although operations are depicted in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order. In certain circumstances, multitasking and parallel processing may be advantageous. Likewise, while several specific implementation details are included in the above discussion, these should not be construed as limiting the scope of the present disclosure. Certain features that are described in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination.
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are example forms of implementing the claims.

Claims (11)

1. An application control method based on virtual reality space is characterized by comprising the following steps:
responding to temporary exit operation of a target application, and switching an application process of the target application into background process operation, wherein when the application process of the target application is switched into background process operation, an application interface of the target application is stopped from being displayed in the virtual reality space;
displaying an application identification model corresponding to the target application in the virtual reality space;
and switching the application process into a foreground process operation in response to a first preset trigger operation on the application identification model, wherein when the application process is switched into the foreground process operation, an application interface of the target application is displayed in the virtual reality space.
2. The method of claim 1, wherein the responding to the temporary exit operation for the target application comprises:
detecting a trigger operation of a preset operation control on the control equipment, wherein the trigger time length in response to the trigger operation is longer than a preset time length; or,
identifying the image information shot by the camera to the user to obtain gesture information of the user,
and responding to the matching of the gesture information of the user and the preset gesture information.
3. The method of claim 1, wherein the displaying in the virtual reality space an application identification model corresponding to the target application comprises:
acquiring the panel position of an operation panel in the virtual reality space;
determining a target display position corresponding to the preset panel position;
the application identification model is displayed at the target display location in the virtual reality space.
4. The method of claim 3, wherein the determining a target display position corresponding to the preset panel position comprises:
determining an associated area of a preset azimuth at the preset panel position;
and determining the target display position in the association area.
5. The method of claim 1, wherein the displaying the application interface of the target application in the virtual reality space comprises:
displaying a preset switching animation in the virtual reality space;
and displaying the application interface after the preset switching animation is displayed.
6. The method as recited in claim 1, further comprising:
identifying the space position of a click mark of preset control equipment;
and under the condition that the space position of the click mark is matched with the space position of a closing control of an application interface of the target application, closing an application process of the target application in response to the click confirmation operation of the closing control, wherein after the application process of the target application is closed, the application interface of the target application is closed.
7. The method as recited in claim 1, further comprising:
and closing the application process of the target application in response to a second preset triggering operation of the application identification model, wherein after the application process of the target application is closed, an application interface of the target application is closed.
8. The method of claim 7, wherein the responding to the first preset trigger operation to the application identification model comprises:
identifying the space position of a click mark of preset control equipment;
responding to clicking operation of a preset operation control on the control equipment under the condition that the clicking mark space position is matched with the space position of the application identification model;
the responding to the second preset triggering operation of the application identification model comprises the following steps:
and under the condition that the click mark space position is matched with the space position of the application identification model, responding to the long-press operation of the preset operation control.
9. An application control device based on virtual reality space, comprising:
the process control module is used for responding to the temporary exit operation of the target application and switching the application process of the target application into background process operation, wherein when the application process of the target application is switched into background process operation, the application interface of the target application is stopped from being displayed in the virtual reality space;
the display module is used for displaying an application identification model corresponding to the target application in the virtual reality space;
the process control module is further configured to switch the application process to a foreground process operation in response to a first preset trigger operation on the application identification model, where when the application process is switched to the foreground process operation, an application interface of the target application is displayed in the virtual reality space.
10. An electronic device, the electronic device comprising:
a processor;
a memory for storing the processor-executable instructions;
the processor is configured to read the executable instructions from the memory and execute the executable instructions to implement the virtual reality space based application control method of any of the preceding claims 1-8.
11. A computer-readable storage medium, characterized in that the computer-readable storage medium stores a computer program for executing the virtual reality space based application control method of any of the preceding claims 1-8.
CN202210964215.XA 2022-08-11 2022-08-11 Application control method, device, equipment and medium based on virtual reality space Pending CN117632391A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210964215.XA CN117632391A (en) 2022-08-11 2022-08-11 Application control method, device, equipment and medium based on virtual reality space

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210964215.XA CN117632391A (en) 2022-08-11 2022-08-11 Application control method, device, equipment and medium based on virtual reality space

Publications (1)

Publication Number Publication Date
CN117632391A true CN117632391A (en) 2024-03-01

Family

ID=90016826

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210964215.XA Pending CN117632391A (en) 2022-08-11 2022-08-11 Application control method, device, equipment and medium based on virtual reality space

Country Status (1)

Country Link
CN (1) CN117632391A (en)

Similar Documents

Publication Publication Date Title
CN112051961A (en) Virtual interaction method and device, electronic equipment and computer readable storage medium
CN114461064B (en) Virtual reality interaction method, device, equipment and storage medium
US20230405475A1 (en) Shooting method, apparatus, device and medium based on virtual reality space
CN111710048B (en) Display method and device and electronic equipment
US20240028130A1 (en) Object movement control method, apparatus, and device
CN115761103A (en) Control method and device based on virtual reality, electronic equipment and storage medium
CN117632391A (en) Application control method, device, equipment and medium based on virtual reality space
CN117319725A (en) Subtitle display method, device, equipment and medium
CN113703704A (en) Interface display method, head-mounted display device and computer readable medium
CN117631810A (en) Operation processing method, device, equipment and medium based on virtual reality space
CN117641025A (en) Model display method, device, equipment and medium based on virtual reality space
US20230377248A1 (en) Display control method and apparatus, terminal, and storage medium
CN117765207A (en) Virtual interface display method, device, equipment and medium
US20240046588A1 (en) Virtual reality-based control method, apparatus, terminal, and storage medium
CN117572994A (en) Virtual object display processing method, device, equipment and medium
CN118244879A (en) Object movement control method, device, equipment and medium
CN117640919A (en) Picture display method, device, equipment and medium based on virtual reality space
CN117636528A (en) Voting processing method, device, equipment and medium based on virtual reality space
CN117376591A (en) Scene switching processing method, device, equipment and medium based on virtual reality
US20240161390A1 (en) Method, apparatus, electronic device and storage medium for control based on extended reality
US20240153211A1 (en) Methods, apparatuses, terminals and storage media for display control based on extended reality
CN117762280A (en) Man-machine interaction method, device and equipment
CN117641026A (en) Model display method, device, equipment and medium based on virtual reality space
CN118484078A (en) Virtual resource processing method, device, equipment and medium based on virtual reality
CN117354484A (en) Shooting processing method, device, equipment and medium based on virtual reality

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination