CN117376591A - Scene switching processing method, device, equipment and medium based on virtual reality - Google Patents

Scene switching processing method, device, equipment and medium based on virtual reality Download PDF

Info

Publication number
CN117376591A
CN117376591A CN202210774820.0A CN202210774820A CN117376591A CN 117376591 A CN117376591 A CN 117376591A CN 202210774820 A CN202210774820 A CN 202210774820A CN 117376591 A CN117376591 A CN 117376591A
Authority
CN
China
Prior art keywords
video stream
seat
switching
virtual reality
playing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210774820.0A
Other languages
Chinese (zh)
Inventor
冀利悦
黄翔宇
赵文珲
贺翔
栾亮勇
尚猛
吴培培
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Zitiao Network Technology Co Ltd
Original Assignee
Beijing Zitiao Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Zitiao Network Technology Co Ltd filed Critical Beijing Zitiao Network Technology Co Ltd
Priority to CN202210774820.0A priority Critical patent/CN117376591A/en
Publication of CN117376591A publication Critical patent/CN117376591A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/23412Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs for generating or manipulating the scene composition of objects, e.g. MPEG-4 objects

Abstract

The embodiment of the disclosure relates to a scene switching processing method, device, equipment and medium based on virtual reality, wherein the method comprises the following steps: playing a first video stream corresponding to a first viewing seat in a virtual reality space; responding to a switching trigger instruction for switching the first viewing seat to the second viewing seat, and rendering the first material to continue playing in the virtual reality space, wherein the playing of the first video stream is stopped when the rendering of the first material is finished; and acquiring a second video stream corresponding to the second watching seat, and playing the second material rendered by the second video stream in the virtual reality space, wherein when the second material is rendered, the scene switching is completed to play the second video stream. In the embodiment of the disclosure, the transition effect during scene switching in the virtual space is realized, and the sense of reality of scene switching in the virtual space is improved.

Description

Scene switching processing method, device, equipment and medium based on virtual reality
Technical Field
The disclosure relates to the technical field of virtual display, and in particular relates to a scene switching processing method, device, equipment and medium based on virtual reality.
Background
Virtual Reality (VR) technology, also known as Virtual environments, moods, or artificial environments, refers to technology that utilizes a computer to generate a Virtual world that can directly impart visual, auditory, and tactile sensations to participants and allow them to interactively observe and operate. The improvement of VR realism to make the experience of virtual reality space and real physical space feel similar becomes a mainstream.
In the related art, viewing of live content such as online concert can be realized based on virtual reality technology, and a user can view a concert such as a real live concert in a virtual space.
However, in the prior art, it is often required to provide a plurality of viewing seats in a virtual reality space, so as to improve the realism of the switching when different viewing seats are switched.
Disclosure of Invention
In order to solve the above technical problems or at least partially solve the above technical problems, the present disclosure provides a method, an apparatus, a device, and a medium for processing scene switching based on virtual reality, which implement a transition effect when switching a scene in a virtual space, and promote a sense of realism of scene switching in the virtual space.
The embodiment of the disclosure provides a scene switching processing method based on virtual reality, which comprises the following steps: playing a first video stream corresponding to a first viewing seat in a virtual reality space; responding to a switching trigger instruction for switching the first viewing seat to the second viewing seat, and rendering a first material to continue playing in the virtual reality space, wherein when the first material rendering is finished, the playing of the first video stream is stopped; and acquiring a second video stream corresponding to the second watching seat, and playing second materials rendered by the second video stream in the virtual reality space, wherein when the second materials are rendered, scene switching is completed to play the second video stream.
The embodiment of the disclosure also provides a scene switching processing device based on virtual reality, which comprises: the playing module is used for playing a first video stream corresponding to the first watching seat in the virtual reality space; the first switching processing module is used for responding to a switching trigger instruction for switching the first viewing seat to the second viewing seat, rendering the first material on the first video stream in the virtual reality space and continuing playing, wherein the playing of the first video stream is stopped when the rendering of the first material is finished; and the second switching processing module is used for acquiring a second video stream corresponding to the second watching seat, and rendering a second material on the second video stream in the virtual reality space for playing, wherein when the second material is rendered, scene switching is completed to play the second video stream.
The embodiment of the disclosure also provides an electronic device, which comprises: a processor; a memory for storing the processor-executable instructions; the processor is configured to read the executable instructions from the memory and execute the instructions to implement the virtual reality-based scene switching processing method according to the embodiments of the present disclosure.
The embodiment of the present disclosure also provides a computer-readable storage medium storing a computer program for executing the virtual reality-based scene switching processing method as provided by the embodiment of the present disclosure.
Compared with the prior art, the technical scheme provided by the embodiment of the disclosure has the following advantages:
according to the scene switching processing scheme based on virtual reality, a first video stream corresponding to a first watching seat is played in a virtual reality space, a switching trigger instruction for switching the first watching seat to a second watching seat is responded, the first video stream is rendered in the virtual reality space to continue playing, the first video stream is stopped being played when the first material is rendered, further, a second video stream corresponding to a second watching seat is obtained, the second material is rendered in the virtual reality space to be played, and scene switching is completed to play the second video stream when the second material is rendered. Therefore, in the embodiment of the disclosure, the transition effect in the scene switching in the virtual space is realized, and the sense of realism of the scene switching in the virtual space is improved.
Drawings
The above and other features, advantages, and aspects of embodiments of the present disclosure will become more apparent by reference to the following detailed description when taken in conjunction with the accompanying drawings. The same or similar reference numbers will be used throughout the drawings to refer to the same or like elements. It should be understood that the figures are schematic and that elements and components are not necessarily drawn to scale.
Fig. 1 is a schematic view of an application scenario of a virtual reality device according to an embodiment of the present disclosure;
FIG. 2 is a schematic illustration of a seat layout for a spectator provided in an embodiment of the present disclosure;
fig. 3 is a schematic flow chart of a scene switching processing method based on virtual reality according to an embodiment of the disclosure;
fig. 4 is a schematic view of a scene based on virtual reality scene switching according to an embodiment of the present disclosure;
fig. 5 is a schematic view of another scene based on virtual reality scene switching according to an embodiment of the disclosure;
fig. 6 is a schematic view of another scene based on virtual reality scene switching according to an embodiment of the disclosure;
fig. 7 is a schematic view of another scene based on virtual reality scene switching according to an embodiment of the disclosure;
fig. 8A is a schematic view of another scene based on virtual reality scene switching according to an embodiment of the disclosure;
Fig. 8B is a schematic view of another scene based on virtual reality scene switching according to an embodiment of the disclosure;
fig. 9 is a schematic structural diagram of a scene switching processing device based on virtual reality according to an embodiment of the disclosure;
fig. 10 is a schematic structural diagram of an electronic device according to an embodiment of the disclosure.
Detailed Description
Embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While certain embodiments of the present disclosure have been shown in the accompanying drawings, it is to be understood that the present disclosure may be embodied in various forms and should not be construed as limited to the embodiments set forth herein, but are provided to provide a more thorough and complete understanding of the present disclosure. It should be understood that the drawings and embodiments of the present disclosure are for illustration purposes only and are not intended to limit the scope of the present disclosure.
It should be understood that the various steps recited in the method embodiments of the present disclosure may be performed in a different order and/or performed in parallel. Furthermore, method embodiments may include additional steps and/or omit performing the illustrated steps. The scope of the present disclosure is not limited in this respect.
The term "including" and variations thereof as used herein are intended to be open-ended, i.e., including, but not limited to. The term "based on" is based at least in part on. The term "one embodiment" means "at least one embodiment"; the term "another embodiment" means "at least one additional embodiment"; the term "some embodiments" means "at least some embodiments. Related definitions of other terms will be given in the description below.
It should be noted that the terms "first," "second," and the like in this disclosure are merely used to distinguish between different devices, modules, or units and are not used to define an order or interdependence of functions performed by the devices, modules, or units.
It should be noted that references to "one", "a plurality" and "a plurality" in this disclosure are intended to be illustrative rather than limiting, and those of ordinary skill in the art will appreciate that "one or more" is intended to be understood as "one or more" unless the context clearly indicates otherwise.
The names of messages or information interacted between the various devices in the embodiments of the present disclosure are for illustrative purposes only and are not intended to limit the scope of such messages or information.
Some technical concepts or noun concepts referred to herein are described in association with:
the virtual reality device, the terminal for realizing the virtual reality effect, may be provided in the form of glasses, a head mounted display (Head Mount Display, HMD), or a contact lens for realizing visual perception and other forms of perception, but the form of the virtual reality device is not limited to this, and may be further miniaturized or enlarged as needed.
The virtual reality device described in the embodiments of the present invention may include, but is not limited to, the following types:
a computer-side virtual reality (PCVR) device performs related computation of a virtual reality function and data output by using a PC side, and an external computer-side virtual reality device realizes a virtual reality effect by using data output by the PC side.
The mobile virtual reality device supports setting up a mobile terminal (such as a smart phone) in various manners (such as a head-mounted display provided with a special card slot), performing related calculation of a virtual reality function by the mobile terminal through connection with the mobile terminal in a wired or wireless manner, and outputting data to the mobile virtual reality device, for example, watching a virtual reality video through an APP of the mobile terminal.
The integrated virtual reality device has a processor for performing the calculation related to the virtual function, and thus has independent virtual reality input and output functions, and is free from connection with a PC or a mobile terminal, and has high degree of freedom in use.
Virtual reality objects, objects that interact in a virtual scene, objects that are stationary, moving, and performing various actions in a virtual scene, such as virtual persons corresponding to a user in a live scene, are controlled by a user or a robot program (e.g., an artificial intelligence based robot program).
As shown in fig. 1, HMDs are relatively light, ergonomically comfortable, and provide high resolution content with low latency. The sensor (such as a nine-axis sensor) for detecting the gesture in the virtual reality device is arranged in the virtual reality device, and is used for detecting the gesture change of the virtual reality device in real time, if the user wears the virtual reality device, when the gesture of the head of the user changes, the real-time gesture of the head is transmitted to the processor, so that the gaze point of the sight of the user in the virtual environment is calculated, an image in the gaze range (namely a virtual view field) of the user in the three-dimensional model of the virtual environment is calculated according to the gaze point, and the image is displayed on the display screen, so that the user looks like watching in the real environment.
In this embodiment, when a user wears the HMD device and opens a predetermined application program, for example, a live video application program, the HMD device may run corresponding virtual scenes, where the virtual scenes may be simulation environments for the real world, semi-simulation virtual scenes, or pure virtual scenes. The virtual scene may be any one of a two-dimensional virtual scene, a 2.5-dimensional virtual scene or a three-dimensional virtual scene, and the dimension of the virtual scene is not limited in the embodiment of the present application. For example, the virtual scene may include characters, sky, land, sea, etc., the land may include environmental elements such as desert, city, etc., the user may control the virtual object to move in the virtual scene, and may also interactively control the controls, models, presentations, characters, etc. in the virtual scene by means of a handle device, a bare hand gesture, etc.
As mentioned above, in the virtual reality space, if the user has a need to switch between different spectator seats, for example, the user is watching the virtual reality space, as shown in fig. 2, a plurality of spectator seats are generally included, and distances and angles of stages of the virtual space watched by the different spectator seats are different, so that the picture information watched by the different spectator seats is similar to the effect achieved by the different seats in the real scene. In the disclosed embodiments, in order to enhance the real experience in virtual reality space, the transition effect is played when switching from the spectator seat to enhance the realism of the transition, thereby enhancing the transition effect.
In order to solve the above-mentioned problems, the embodiments of the present disclosure provide a scene switching processing method based on virtual reality, and the method is described below with reference to specific embodiments.
Fig. 3 is a flow chart of a virtual reality-based scene switching processing method according to an embodiment of the disclosure, where the method may be performed by a virtual reality-based scene switching processing device, and the device may be implemented by software and/or hardware, and may be generally integrated in an electronic device. As shown in fig. 3, the method includes:
Step 301, playing a first video stream corresponding to a first viewing seat in a virtual reality space.
The visual elements such as the video stream and the rendering effect observed in the embodiment of the disclosure are watched by a user wearing the virtual reality device.
In one embodiment of the present disclosure, if the viewing position is located at the first viewing seat, a first video stream corresponding to the first viewing seat is played in the virtual reality space.
In an embodiment of the present disclosure, an angle and a distance of the first viewing seat from the preset virtual stage scene may be determined, a field of view range is determined based on the angle and the distance, and a stage video stream within the field of view range is determined to be the first video stream. The field of view range in this embodiment refers to a range in which a user is required to photograph a virtual reality scene during viewing of VR video, and for this embodiment, parameters related to controlling the photographing range of the virtual reality device, such as a field of view (FOV), may be preset. The field of view range can be adjusted according to the requirements of users, and then required photos or videos and the like are shot.
Step 302, in response to a switching trigger instruction for switching the first viewing seat to the second viewing seat, rendering the first material on the first video stream in the virtual reality space for continuous playing, wherein playing of the first video stream is stopped when the rendering of the first material is finished.
It should be noted that, the acquisition modes of the switching trigger instruction in different application scenarios are different, and examples are as follows:
in one embodiment of the disclosure, a seat switching control is preset on the virtual reality device, in this real time, whether the seat switching control on the virtual reality device is opened is detected, and if the seat switching control is opened, a switching trigger instruction for switching a first viewing seat to a second viewing seat is responded.
In one embodiment of the present disclosure, a view mat bitmap is displayed in a virtual reality space, wherein image information captured by a camera on a user may be monitored, then, whether a preset condition for displaying the view mat bitmap is met or not is judged according to a user hand or a user hand-held device (such as a handle) in the image information, if the preset condition for displaying the view mat bitmap is judged to be met, the view mat bitmap is displayed in the virtual reality space, and finally, the corresponding view mat bitmap is displayed by identifying action information of the user hand or the user hand-held device.
For example, a camera may be used to capture an image of a user's hand or an image of a user's handheld device, and determine a user's hand gesture or a change in position of the handheld device in the image based on an image recognition technique, and if it is determined that the user's hand or the user's handheld device lifts up by a certain extent so that the user's virtual hand or the virtual handheld device mapped in the virtual reality space enters into a current viewing angle range of the user, a viewing seat bitmap may be displayed in the virtual reality space, where the viewing seat bitmap marks the current viewing seat for the first viewing seat, detects whether to trigger the second viewing seat, and if the second viewing seat is triggered, responds to a switch trigger instruction for switching the first viewing seat to the second viewing seat. For example, as shown in fig. 4, when a trigger operation (a trigger operation performed by a "ray" in the figure) performed by the user on the second viewing seat b in the viewing seat map is detected, a switching trigger instruction for switching the first viewing seat to the second viewing seat is responded.
In one embodiment of the present disclosure, a trigger action to switch to a second viewing seat is detected, and if the trigger action is present, a switch trigger instruction to switch the first viewing seat to the second viewing seat is responded.
In this optional mode, the image information shot by the camera on the user can be monitored, then whether the triggering operation of switching to the second watching seat is met or not is judged according to the user hand or the user handheld device (such as a handle) in the image information, if the triggering operation corresponding to the switching triggering instruction is judged to be met, the switching triggering instruction is obtained, and finally the switching triggering instruction for switching to the second watching seat is obtained by identifying the action information of the user hand or the user handheld device.
For example, a camera may be used to capture an image of a user's hand or an image of a user's handheld device, and based on an image recognition technique, determine a user's hand gesture or a change in a position of the handheld device in the image, and if it is determined that the user's hand or the user's handheld device is lifted by a certain extent, and a trigger action is performed according to a preset gesture track, obtain a switch trigger instruction for switching to the second viewing seat.
Further, after the switching trigger instruction is acquired, the first material is rendered for the first video stream in the virtual reality space to continue playing in response to the switching trigger instruction for switching the first viewing seat to the second viewing seat.
The first material is used for prompting a watching user about to switch scenes, and when the first material is rendered, the first video stream is stopped being played. Therefore, in this embodiment, the first material is rendered to the first video stream in the virtual reality space for further playing, so as to achieve an effect similar to the transition in the reality space, and improve the sense of realism of scene switching.
It is to be readily appreciated that the first material may be any material that may visually embody the switching of the first video stream, including, but not limited to, one or more of sound, animation, text, patterns, colors, etc.
Step 303, obtaining a second video stream corresponding to the second watching seat, and playing the second material rendered by the second video stream in the virtual reality space, wherein when the second material is rendered, scene switching is completed to play the second video stream.
In one embodiment of the present disclosure, as mentioned above, the angles and distances and the like at which the first viewing seat and the second viewing seat are viewed are different, and therefore, the video streams on the stage that are viewed, that is, the second video streams, are necessarily different, and therefore, in the present embodiment, the second video streams corresponding to the second viewing seat are acquired.
In an embodiment of the present disclosure, an angle and a distance of the second viewing seat from the preset virtual stage scene may be determined, a field of view range is determined based on the angle and the distance, and a stage video stream within the field of view range is determined to be the second video stream. The field of view range in this embodiment refers to a range in which a user is required to photograph a virtual reality scene during viewing of VR video, and for this embodiment, parameters related to controlling the photographing range of the virtual reality device, such as a field of view (FOV), may be preset. The field of view range can be adjusted according to the requirements of users, and then required photos or videos and the like are shot.
And after the second video stream corresponding to the second watching seat is obtained, playing the second material rendered by the second video stream in the virtual reality space, wherein when the second material is rendered, the scene is switched to complete playing the second video stream.
It is readily understood that the second material may be any material that may visually embody the switching of the second video stream, including, but not limited to, one or more of sound, animation, text, patterns, colors, etc.
In this embodiment, the second material is played in the virtual reality space by rendering the second video stream, so that transition display of switching the first video stream to the second video stream is realized, a switching transition effect in the case of machine position switching in reality is realized, and the reality of transition is improved.
Therefore, in the scene switching processing process based on virtual reality, as shown in fig. 5, a first video stream corresponding to a first watching seat is played in a virtual reality space, a first material is rendered for the first video stream in the virtual reality space to continue playing in response to a switching trigger instruction for switching the first watching seat to a second watching seat, and on the other hand, the switching reality is promoted on the one hand, and on the other hand, because time is required for loading a second video stream, the problem of switching vision clamping is avoided.
And further, a second video stream corresponding to the second watching seat is obtained, and second materials are rendered on the second video stream in the virtual reality space for playing, wherein the second materials are used for prompting the watching user to switch scenes, and the second video stream is played after the scene switching is completed when the special effect rendering is finished, wherein the second materials are played on the second video stream in the virtual reality space, so that the effect of 'gradual conversion' of the machine position is formed, the realism of the scene switching is improved, the sense of hardness of the video stream switching is avoided, and the fluency of the switching is improved.
In summary, according to the scene switching processing method based on virtual reality provided by the embodiment of the disclosure, a first video stream corresponding to a first viewing seat is played in a virtual reality space, a switching trigger instruction for switching the first viewing seat to a second viewing seat is responded, a first material is rendered for continuing playing in the virtual reality space, wherein when the first material is rendered, the first video stream is stopped from being played, further, a second video stream corresponding to a second viewing seat is obtained, and a second material is rendered for playing in the virtual reality space, wherein when the second material is rendered, scene switching is completed to play the second video stream. Therefore, in the embodiment of the disclosure, the transition effect in the scene switching in the virtual space is realized, and the sense of realism of the scene switching in the virtual space is improved.
Based on the above embodiment, the first material and the second material may be preset, or may be flexibly matched according to the scene. If the first material and the second material are flexibly matched according to the scenes, different modes can be selected to determine the first material and the second material according to different scenes. The first material may correspond to the second material to ensure the connection of the switching, for example, the first material corresponds to the "open eye" special effect, and the second material corresponds to the "closed eye" special effect.
In some possible embodiments, a target position of a target object in the first video stream is determined, where the target object may be understood as a playing object corresponding to the video stream, for example, may be a stage performer, etc., and the target position may be understood as a coordinate position of the target object in the virtual reality space.
In this embodiment, according to a first distance from the first viewing seat to the target position, a first material matched with the first video stream is determined, for example, according to a preset corresponding relationship corresponding to the first viewing seat, a first material corresponding to the first distance is queried. The first material and the first distance from the first watching seat to the target position are corresponding, so that the watching atmosphere sense can be improved.
For example, when the target object is a stage performer, the closer the first distance is, the stronger the visual impact of the corresponding first material (for example, the visual impact of the special effect material is reflected by the sound size, the animation playing frequency, and the like), and the farther the first distance is, the weaker the visual impact of the corresponding first material is, and the like. The effect of this and the actual stage of watching is unanimous, has further promoted the sense of realism.
Similarly, in this embodiment, according to the second distance from the second viewing seat to the target position, the second material matched with the second video stream is determined, for example, according to the preset corresponding relationship corresponding to the second viewing seat, the second material corresponding to the second distance is queried. The second distance between the second material and the target position of the second watching seat is corresponding, so that the watching atmosphere can be improved, for example, when the target object is a stage performer, the closer the second distance is, the stronger the visual impact of the corresponding second material is (for example, the visual impact of the special effect material is reflected by the sound size, the animation playing frequency and the like), and the farther the second distance is, the weaker the visual impact of the corresponding second material is. The effect of this and the actual stage of watching is unanimous, has further promoted the sense of realism.
In some possible embodiments, the first material is a closed-eye simulation material, and the second material is an open-eye simulation material, where the closed-eye simulation material is any visual presentation of a "closed-eye" special effect material, and the open-eye simulation material is any visual presentation of a "open-eye" special effect material, so that when a stage performance is watched, a user can switch from a "gradually invisible" first video stream of the first audience to a "gradually visible" second video stream of the second audience in a display scene through the closed-eye simulation material and the open-eye simulation material.
In this embodiment, when the first material is rendered for the first video stream and continues to be played, the process of gradually performing from clear to blurred on the first video stream from the periphery to the middle is performed, and the sound of the first video stream is gradually reduced, where the process of gradually performing from clear to blurred on the first video stream from the periphery to the middle may be as shown in fig. 6, and the viewing effect that the user gradually walks away from the first audience in real space is simulated (the implementation shown in the figure is merely an example, and in fact, the "eye-closing" special effect may be implemented by gradually reducing the brightness of the first video stream in addition to the implementation shown in fig. 6, which is not listed here).
In this embodiment, rendering the second material for the second video stream is performed as follows: the process of gradually blurring the second video stream from middle to periphery and gradually increasing the sound of the second video stream, where the process of gradually blurring the second video stream from middle to periphery may continue as shown in fig. 6, and the viewing effect of gradually approaching the second auditorium by the user in real space is simulated (the implementation shown in the figure is merely an example, in fact, besides the implementation shown in fig. 6, the effect of "opening eyes" may be implemented by gradually increasing the brightness of the second video stream, which is not specifically shown herein).
In some possible embodiments, the first material is a countdown closing analog material, and the second material is a countdown starting analog material, where the countdown closing analog material may be any material that achieves a countdown effect, for example, may be a stopwatch countdown animation material, etc., and the countdown starting analog material may be any material that achieves a countdown effect, for example, may be a stopwatch countdown animation material, etc.
In this embodiment, the countdown animation identifier is rendered for the first video stream, and the sound of the first video stream is gradually reduced, where, for example, as shown in fig. 7, the countdown animation identifier is rendered for the first video stream (in the figure, the countdown animation is shown), and after the countdown is finished, the second material is rendered for the second video stream, that is, the countdown animation identifier is rendered for the second video stream, and the sound of the second video stream is gradually increased, where, for example, as shown in fig. 7, the countdown animation identifier is continuously rendered for the second video stream (in the figure, the countdown animation is shown). Therefore, switching from the countdown to the countdown is intuitively reflected from the first audience to the second audience, display abrupt sense caused by suddenly switching the displayed video stream is avoided, a transition stage of scene transition in reality is simulated, and the sense of reality of the scene transition is improved.
In addition, in order to further improve the sense of reality of switching, when the first material rendering is finished, the first video stream is stopped being played, and a scene switching waiting picture is displayed in the virtual reality space. Therefore, for example, when the loading time is long, that is, after the first video stream stops playing, the second video stream is not loaded yet, the waiting picture can be switched through the display scene, so that the problem of display blocking is avoided, and the display effect is improved.
In some possible embodiments, a waiting identifier is added to a preset target image as a scene switching waiting screen, where any preset adding waiting identifier may be included in the scene switching waiting screen, for example, the adding waiting identifier may be a "… …" identifier on the target image, so as to identify that a waiting is currently being rendered, and further, load a second video stream corresponding to the second viewing seat.
In some possible embodiments, the gaussian blur processing is performed on the corresponding video frame when the first video stream stops playing, and the image after the gaussian blur processing is used as a scene switching waiting picture.
In some possible embodiments, the black screen image is used as a scene-switching waiting screen.
In this embodiment, whether the loading time of the second video stream is greater than a preset threshold is detected, if the loading time is detected to be greater than the preset threshold, as shown in fig. 8A, when the rendering of the first material is finished, the playing of the first video stream is stopped, a scene switching waiting picture is displayed in the virtual reality space, and the second video stream corresponding to the second viewing seat is loaded, where, as shown in fig. 8B, a loading identifier is displayed in the scene switching waiting picture (black screen in the drawing) on the basis of the one shown in fig. 7, and the loading identifier may be an identifier such as … … so as to intuitively embody that the present stage is in the loading waiting.
In order to further improve the display effect, the loading identifier may be further generated according to feature information associated with the first video stream or the second video stream, where the associated feature information may be a target object type feature corresponding to the video stream, when the target object is a singer, the corresponding loading identifier may be a "music" symbol, and when the target object is a food, the corresponding loading identifier may be a "food" symbol, and so on. The associated feature information may be a color combination feature corresponding to the video stream (the first video stream or the second video stream), for example, when the color combination feature corresponding to the video stream is a red-blue color system, the corresponding loading identifier may be selected from an identifier library corresponding to the "red-blue color system", or the like.
In order to further improve the display effect, after the scene switching waiting picture is displayed in the virtual reality space, the scene switching sound effect can be played until the second video stream is loaded, wherein the scene switching sound effect can be obtained by matching according to the type of the video stream, can also be set by a user according to personal preference, and the like. In this embodiment, the scene change sound effect is played to further enhance the scene change effect.
In summary, according to the virtual reality-based scene switching processing method disclosed by the embodiment of the disclosure, the first material and the second material can be flexibly selected according to the scene requirement, so that the effect of switching from the first video stream to the second video stream is improved, and the visual sense realism when the audience is switched in the virtual space scene is further improved.
In order to achieve the above embodiments, the present disclosure further provides a scene switching processing device based on virtual reality. Fig. 9 is a schematic structural diagram of a scene switching processing device based on virtual reality according to an embodiment of the disclosure, where the device may be implemented by software and/or hardware, and may be generally integrated in an electronic device. As shown in fig. 9, the apparatus includes: a play module 910, a first switch processing module 920, a second switch processing module 930, wherein,
a playing module 910, configured to play a first video stream corresponding to a first viewing seat in a virtual reality space;
a first switching processing module 920, configured to respond to a switching trigger instruction for switching the first viewing seat to the second viewing seat, and render, in the virtual reality space, a first material to continue playing, where playing of the first video stream is stopped when the rendering of the first material is finished;
And the second switching processing module 930 is configured to obtain a second video stream corresponding to the second viewing seat, and play the second material rendered by the second video stream in the virtual reality space, where when the rendering of the second material is finished, the scene switching is completed to play the second video stream.
The virtual reality-based scene switching processing device provided by the embodiment of the disclosure can execute the virtual reality-based scene switching processing method provided by any embodiment of the disclosure, and has the corresponding functional modules and beneficial effects of the execution method.
To achieve the above embodiments, the present disclosure also proposes a computer program product comprising a computer program/instruction which, when executed by a processor, implements the virtual reality based scene cut processing method in the above embodiments.
Fig. 10 is a schematic structural diagram of an electronic device according to an embodiment of the disclosure.
Referring now in particular to fig. 10, a schematic diagram of an electronic device 1000 suitable for use in implementing embodiments of the present disclosure is shown. The electronic device 1000 in the embodiments of the present disclosure may include, but is not limited to, mobile terminals such as mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablet computers), PMPs (portable multimedia players), in-vehicle terminals (e.g., in-vehicle navigation terminals), and the like, and stationary terminals such as digital TVs, desktop computers, and the like. The electronic device shown in fig. 10 is merely an example and should not be construed to limit the functionality and scope of use of the disclosed embodiments.
As shown in fig. 10, the electronic device 1000 may include a processing means (e.g., a central processing unit, a graphics processor, etc.) 1001 that may perform various appropriate actions and processes according to a program stored in a Read Only Memory (ROM) 1002 or a program loaded from a storage means 1008 into a Random Access Memory (RAM) 1003. In the RAM 1003, various programs and data necessary for the operation of the electronic apparatus 1000 are also stored. The processing device 1001, the ROM 1002, and the RAM 1003 are connected to each other by a bus 1004. An input/output (I/O) interface 1005 is also connected to bus 1004.
In general, the following devices may be connected to the I/O interface 1005: input devices 1006 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, and the like; an output device 1007 including, for example, a Liquid Crystal Display (LCD), speaker, vibrator, etc.; storage 1008 including, for example, magnetic tape, hard disk, etc.; and communication means 1009. The communication means 1009 may allow the electronic device 1000 to communicate wirelessly or by wire with other devices to exchange data. While fig. 10 shows an electronic device 1000 having various means, it is to be understood that not all of the illustrated means are required to be implemented or provided. More or fewer devices may be implemented or provided instead.
In particular, according to embodiments of the present disclosure, the processes described above with reference to flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a non-transitory computer readable medium, the computer program comprising program code for performing the method shown in the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network via the communication device 1009, or installed from the storage device 1008, or installed from the ROM 1002. When the computer program is executed by the processing apparatus 1001, the above-described functions defined in the virtual reality-based scene change processing method of the embodiment of the present disclosure are executed.
It should be noted that the computer readable medium described in the present disclosure may be a computer readable signal medium or a computer readable storage medium, or any combination of the two. The computer readable storage medium can be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the foregoing. More specific examples of the computer-readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this disclosure, a computer-readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In the present disclosure, however, the computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave, with the computer-readable program code embodied therein. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination of the foregoing. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: electrical wires, fiber optic cables, RF (radio frequency), and the like, or any suitable combination of the foregoing.
In some implementations, the clients, servers may communicate using any currently known or future developed network protocol, such as HTTP (HyperText Transfer Protocol ), and may be interconnected with any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include a local area network ("LAN"), a wide area network ("WAN"), the internet (e.g., the internet), and peer-to-peer networks (e.g., ad hoc peer-to-peer networks), as well as any currently known or future developed networks.
The computer readable medium may be contained in the electronic device; or may exist alone without being incorporated into the electronic device.
The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to: and playing the first video stream corresponding to the first watching seat in the virtual reality space, responding to a switching trigger instruction for switching the first watching seat to the second watching seat, and rendering the first material in the virtual reality space to continue playing, wherein when the first material is rendered, stopping playing the first video stream, further obtaining the second video stream corresponding to the second watching seat, and playing the second material in the virtual reality space, wherein when the second material is rendered, scene switching is completed to play the second video stream. Therefore, in the embodiment of the disclosure, the transition effect in the scene switching in the virtual space is realized, and the sense of realism of the scene switching in the virtual space is improved. Computer program code for carrying out operations of the present disclosure may be written in one or more programming languages, including, but not limited to, an object oriented programming language such as Java, smalltalk, C ++ and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computer (for example, through the Internet using an Internet service provider).
The flowcharts and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units involved in the embodiments of the present disclosure may be implemented by means of software, or may be implemented by means of hardware. Wherein the names of the units do not constitute a limitation of the units themselves in some cases.
The functions described above herein may be performed, at least in part, by one or more hardware logic components. For example, without limitation, exemplary types of hardware logic components that may be used include: a Field Programmable Gate Array (FPGA), an Application Specific Integrated Circuit (ASIC), an Application Specific Standard Product (ASSP), a system on a chip (SOC), a Complex Programmable Logic Device (CPLD), and the like.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. The machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
The foregoing description is only of the preferred embodiments of the present disclosure and description of the principles of the technology being employed. It will be appreciated by persons skilled in the art that the scope of the disclosure referred to in this disclosure is not limited to the specific combinations of features described above, but also covers other embodiments which may be formed by any combination of features described above or equivalents thereof without departing from the spirit of the disclosure. Such as those described above, are mutually substituted with the technical features having similar functions disclosed in the present disclosure (but not limited thereto).
Moreover, although operations are depicted in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order. In certain circumstances, multitasking and parallel processing may be advantageous. Likewise, while several specific implementation details are included in the above discussion, these should not be construed as limiting the scope of the present disclosure. Certain features that are described in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination.
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are example forms of implementing the claims.

Claims (13)

1. The scene switching processing method based on virtual reality is characterized by comprising the following steps of:
playing a first video stream corresponding to a first viewing seat in a virtual reality space;
responding to a switching trigger instruction for switching the first viewing seat to the second viewing seat, and rendering a first material to continue playing in the virtual reality space, wherein when the first material rendering is finished, the playing of the first video stream is stopped;
and acquiring a second video stream corresponding to the second watching seat, and playing second materials rendered by the second video stream in the virtual reality space, wherein when the second materials are rendered, scene switching is completed to play the second video stream.
2. The method of claim 1, wherein the responding to a switch trigger instruction to switch the first viewing seat to a second viewing seat comprises:
Detecting whether a seat switching control on the virtual reality equipment is started, and if the seat switching control is started, responding to a switching trigger instruction for switching the first watching seat to a second watching seat; or,
displaying a view seat bitmap in the virtual reality space, wherein the view seat bitmap marks the current view seat of the first view seat, detecting whether a trigger operation is performed on a second view seat, and if the trigger operation is performed on the second view seat, responding to a switch trigger instruction for switching the first view seat to the second view seat; or,
detecting whether a trigger action for switching to the second watching seat exists or not, and if the trigger action exists, responding to a switching trigger instruction for switching the first watching seat to the second watching seat.
3. The method as recited in claim 1, further comprising:
determining a target position of a target object in the first video stream;
determining a first material matched with the first video stream according to a first distance from the first watching seat to the target position;
and determining a second material matched with the second video stream according to a second distance from the second watching seat to the target position.
4. The method of claim 1, wherein the first material is closed-eye simulated material and the second material is open-eye simulated material,
the rendering of the first material to the first video stream continues to play, including:
gradually carrying out a clear-to-blur processing process on the first video stream from the periphery to the middle, and gradually reducing the sound of the first video stream;
the playing the second material rendered by the second video stream comprises the following steps:
and gradually carrying out a process from blurring to sharpness on the second video stream from the middle to the periphery, and gradually increasing the sound of the second video stream.
5. The method of claim 1, wherein the first material is a countdown shutdown analog material and the second material is a countdown startup analog material,
the rendering of the first material to the first video stream continues to play, including:
rendering the first video stream, closing the animation mark by countdown, and gradually reducing the sound of the first video stream;
the playing the second material rendered by the second video stream comprises the following steps:
and starting the animation identification when the second video stream is rendered and timing, and gradually increasing the sound of the second video stream.
6. The method of claim 1, wherein the obtaining a second video stream corresponding to the second viewing seat comprises:
stopping playing the first video stream when the first material rendering is finished, and displaying a scene switching waiting picture in the virtual reality space;
and loading a second video stream corresponding to the second watching seat.
7. The method of claim 6, wherein displaying a scene cut wait screen in the virtual reality space comprises:
carrying out Gaussian blur processing on a video frame corresponding to the first video stream when the first video stream stops playing, and taking an image subjected to Gaussian blur processing as the scene switching waiting picture; or,
taking the black screen image as the scene switching waiting picture; or,
and adding a waiting mark on a preset target image to serve as the scene switching waiting picture.
8. The method of claim 6, wherein loading a second video stream corresponding to the second viewing seat comprises:
detecting whether the loading time of the second video stream is greater than a preset threshold value;
and if the loading time is detected and known to be larger than the preset threshold value, displaying a loading identifier in the scene switching waiting picture.
9. The method as recited in claim 8, further comprising:
and generating the loading identification according to the characteristic information associated with the first video stream or the second video stream.
10. The method of claim 6, further comprising, after the virtual reality space displays a scene cut wait screen:
and switching the sound effect of the playing scene until the second video stream is loaded.
11. A virtual reality-based scene switching processing apparatus, comprising:
the playing module is used for playing a first video stream corresponding to the first watching seat in the virtual reality space;
the first switching processing module is used for responding to a switching trigger instruction for switching the first viewing seat to the second viewing seat, rendering the first material on the first video stream in the virtual reality space and continuing playing, wherein the playing of the first video stream is stopped when the rendering of the first material is finished;
and the second switching processing module is used for acquiring a second video stream corresponding to the second watching seat, and rendering a second material on the second video stream in the virtual reality space for playing, wherein when the second material is rendered, scene switching is completed to play the second video stream.
12. An electronic device, the electronic device comprising:
a processor;
a memory for storing the processor-executable instructions;
the processor is configured to read the executable instructions from the memory and execute the instructions to implement the virtual reality-based scene cut processing method according to any of claims 1-10.
13. A computer-readable storage medium, characterized in that the storage medium stores a computer program for executing the virtual reality based scene cut processing method of any of the preceding claims 1-10.
CN202210774820.0A 2022-07-01 2022-07-01 Scene switching processing method, device, equipment and medium based on virtual reality Pending CN117376591A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210774820.0A CN117376591A (en) 2022-07-01 2022-07-01 Scene switching processing method, device, equipment and medium based on virtual reality

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210774820.0A CN117376591A (en) 2022-07-01 2022-07-01 Scene switching processing method, device, equipment and medium based on virtual reality

Publications (1)

Publication Number Publication Date
CN117376591A true CN117376591A (en) 2024-01-09

Family

ID=89391622

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210774820.0A Pending CN117376591A (en) 2022-07-01 2022-07-01 Scene switching processing method, device, equipment and medium based on virtual reality

Country Status (1)

Country Link
CN (1) CN117376591A (en)

Similar Documents

Publication Publication Date Title
CN106716302B (en) Method, apparatus, and computer-readable medium for displaying image
CN111701238A (en) Virtual picture volume display method, device, equipment and storage medium
CN112105983B (en) Enhanced visual ability
CN110662105A (en) Animation file generation method and device and storage medium
CN110807769B (en) Image display control method and device
CN112581571A (en) Control method and device of virtual image model, electronic equipment and storage medium
US20220189433A1 (en) Application programming interface for setting the prominence of user interface elements
CN113194329B (en) Live interaction method, device, terminal and storage medium
CN114415907A (en) Media resource display method, device, equipment and storage medium
CN117376591A (en) Scene switching processing method, device, equipment and medium based on virtual reality
CN106775245B (en) User attribute setting method and device based on virtual reality
US20230405475A1 (en) Shooting method, apparatus, device and medium based on virtual reality space
US20240078734A1 (en) Information interaction method and apparatus, electronic device and storage medium
CN117354484A (en) Shooting processing method, device, equipment and medium based on virtual reality
US20230377248A1 (en) Display control method and apparatus, terminal, and storage medium
CN117572994A (en) Virtual object display processing method, device, equipment and medium
WO2023231666A1 (en) Information exchange method and apparatus, and electronic device and storage medium
CN116206090A (en) Shooting method, device, equipment and medium based on virtual reality space
CN117631810A (en) Operation processing method, device, equipment and medium based on virtual reality space
CN117640919A (en) Picture display method, device, equipment and medium based on virtual reality space
CN117641026A (en) Model display method, device, equipment and medium based on virtual reality space
CN117934769A (en) Image display method, device, electronic equipment and storage medium
CN117631904A (en) Information interaction method, device, electronic equipment and storage medium
CN117765207A (en) Virtual interface display method, device, equipment and medium
CN117435041A (en) Information interaction method, device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination