WO2021228200A1 - Method for realizing interaction in three-dimensional space scene, apparatus and device - Google Patents

Method for realizing interaction in three-dimensional space scene, apparatus and device Download PDF

Info

Publication number
WO2021228200A1
WO2021228200A1 PCT/CN2021/093628 CN2021093628W WO2021228200A1 WO 2021228200 A1 WO2021228200 A1 WO 2021228200A1 CN 2021093628 W CN2021093628 W CN 2021093628W WO 2021228200 A1 WO2021228200 A1 WO 2021228200A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
dimensional model
user terminal
pixel
information
Prior art date
Application number
PCT/CN2021/093628
Other languages
French (fr)
Chinese (zh)
Inventor
白杰
姚锟
贾松林
郑深圳
张蕾
Original Assignee
贝壳技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from CN202010401813.7A external-priority patent/CN111562845B/en
Priority claimed from CN202010698810.4A external-priority patent/CN111885398B/en
Application filed by 贝壳技术有限公司 filed Critical 贝壳技术有限公司
Publication of WO2021228200A1 publication Critical patent/WO2021228200A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/239Interfacing the upstream path of the transmission network, e.g. prioritizing client content requests

Definitions

  • the present disclosure relates to virtual reality panoramic technology and streaming media technology, and in particular to a method for realizing three-dimensional space scene interaction, a device for realizing three-dimensional space scene interaction, a storage medium, and electronic equipment.
  • VR panoramic technology is an emerging rich media technology. Because VR panoramic technology can present users with three-dimensional space scenes without blind angles at 720 degrees, and bring users an immersive visual experience, VR panoramic technology is widely used in various fields such as online shopping malls, travel services, and real estate services. How to enable VR panoramic technology to bring users a richer experience is a technical issue worthy of attention.
  • three-dimensional models can give people a stronger visual perception.
  • any view of the object can be presented to the user, and the correct projection relationship can be maintained between the views.
  • the user terminal while the user terminal is presenting a three-dimensional model, it can support real-time voice on-screen interaction between user terminals, that is, in the process of the user terminal presenting the three-dimensional model, the voice of the opposite user of the user terminal can be transmitted in real time To the user terminal, and the voice acquired by the user terminal can also be transmitted to the opposite terminal in real time.
  • a method for realizing interaction in a three-dimensional space scene including: in response to detecting a user operation of setting footprint information in the three-dimensional space scene, determining the user's current position in the three-dimensional space scene The first pixel in the current view corresponding to the perspective; determine the three-dimensional model corresponding to the first pixel; determine the position of the user's footprint information in the three-dimensional model, where the footprint information is used to display when the three-dimensional space scene is viewed ; And set the user's footprint information at the location.
  • an interaction method based on a three-dimensional model including: at a first user terminal presenting a user interface: in response to detecting a user's target interaction operation on the user interface, The server that provides page data in the user interface sends an interaction request for the target interaction operation, where the user interface is used to present a three-dimensional model, and the three-dimensional model establishes an association relationship with the user account logged in the second user terminal in advance; Streaming video obtained by the terminal; and presenting the streaming video and three-dimensional model on the user interface.
  • an interaction method based on a three-dimensional model including: at a second user terminal: in response to receiving an interaction request sent by a server, acquiring a streaming video, wherein the interaction request indicates The first user terminal detects the user's target interaction operation on the user interface presented by the first user terminal, the user interface is used to present a three-dimensional model, and the three-dimensional model establishes a pre-association relationship with the user account logged in by the second user terminal; and sends a stream to the server Media video, where the server is used to send the streaming video to the first user terminal, so that the first user terminal presents the streaming video and the three-dimensional model on the user interface.
  • a device for realizing interaction of a three-dimensional space scene including: a device for executing the method described in any one of the above methods.
  • an interaction device based on a three-dimensional model which is provided in a first user terminal, and the device includes: a device for executing the method described in any one of the above methods.
  • an interaction device based on a three-dimensional model which is provided in a second user terminal, and the device includes: a device for executing the method described in any one of the above methods.
  • an interactive system based on a three-dimensional model, including: a first user terminal for presenting a user interface; a second user terminal; and a server.
  • the second user terminal is in communication connection.
  • the first user terminal is configured to: in response to detecting the user's target interaction operation on the user interface, send an interaction request for the target interaction operation to the server, and the user interface is used to present the three-dimensional model, the three-dimensional model and the second 2.
  • the user account logged in by the user terminal establishes an association relationship in advance; the second user terminal is configured to: obtain the streaming video; and send the streaming video to the server; the server is configured to: send the streaming media to the first user terminal Video; and the first user terminal is configured to: present a streaming video and a three-dimensional model on the user interface.
  • a non-transitory computer-readable storage medium stores a computer program. method.
  • an electronic device including: a processor; and a memory for storing processor-executable instructions.
  • the processor-executable instructions implement any of the above methods when executed by the processor. The method described in one item.
  • a computer program product including a computer program, which when executed by a computer causes the computer to implement the method described in any one of the above methods.
  • FIG. 1 is a schematic diagram of an embodiment of an applicable scenario of the present disclosure
  • FIG. 2 is a flowchart of an embodiment of a method for realizing interaction of a three-dimensional space scene of the present disclosure
  • FIG. 3 is a flowchart of an embodiment of determining a three-dimensional model corresponding to a first pixel in the present disclosure
  • FIG. 4 is a flowchart of another embodiment of determining a three-dimensional model corresponding to a first pixel point of the present disclosure
  • FIG. 5 is a flowchart of an embodiment of presenting footprint information for browsing users in the present disclosure
  • FIG. 6 is a schematic structural diagram of an embodiment of an apparatus for realizing interaction in a three-dimensional space scene of the present disclosure
  • FIG. 7 is a flowchart of an embodiment of the first interaction method based on a three-dimensional model of the present disclosure.
  • 8A-8C are schematic diagrams of application scenarios for the embodiment of FIG. 7.
  • Fig. 9 is a flowchart of another embodiment of the first three-dimensional model-based interaction method of the present disclosure.
  • FIG. 10 is a flowchart of another embodiment of the first interaction method based on a three-dimensional model of the present disclosure.
  • FIG. 11 is a flowchart of an embodiment of the second three-dimensional model-based interaction method of the present disclosure.
  • FIG. 12 is a flowchart of another embodiment of the second three-dimensional model-based interaction method of the present disclosure.
  • FIG. 13 is a flowchart of an embodiment of the first interactive device based on a three-dimensional model of the present disclosure.
  • FIG. 14 is a flowchart of an embodiment of the second interactive device based on a three-dimensional model of the present disclosure.
  • FIG. 15 is a schematic diagram of interaction of an embodiment of the interactive system based on a three-dimensional model of the present disclosure.
  • Fig. 16 is a structural diagram of an electronic device provided by an exemplary embodiment of the present disclosure.
  • plural may refer to two or more than two, and “at least one” may refer to one, two, or more than two.
  • the term "and/or" in the present disclosure is only an association relationship that describes associated objects, indicating that there can be three relationships, such as A and/or B, which can mean: A alone exists, and A and B exist at the same time. There are three cases of B alone.
  • the character "/" in the present disclosure generally indicates that the associated objects before and after are in an "or" relationship.
  • the embodiments of the present disclosure can be applied to electronic devices such as terminal devices, computer systems, servers, etc., which can operate with many other general-purpose or special-purpose computing system environments or configurations.
  • Examples of well-known terminal devices, computing systems, environments and/or configurations suitable for use with electronic devices such as terminal devices, computer systems, or servers include, but are not limited to: personal computer systems, server computer systems, thin clients, and thick clients Computers, handheld or laptop devices, microprocessor-based systems, set-top boxes, programmable consumer electronics, network personal computers, small computer systems, large computer systems, and distributed cloud computing technology environments including any of the above systems, etc.
  • Electronic devices such as terminal devices, computer systems, and servers can be described in the general context of computer system executable instructions (such as program modules) executed by the computer system.
  • program modules may include routines, programs, object programs, components, logic, data structures, etc., which perform specific tasks or implement specific abstract data types.
  • the computer system/server can be implemented in a distributed cloud computing environment.
  • tasks can be performed by remote processing devices linked through a communication network.
  • program modules may be located on a storage medium of a local or remote computing system including a storage device.
  • the inventor found that in the process of a user experiencing a three-dimensional space scene by adjusting his current perspective, some feelings such as emotions and thoughts are often generated. If the user can set the footprint information used to characterize his feelings into the three-dimensional space scene, it will not only help improve the user’s sense of participation, but the footprint information left by the user can also provide other users watching the three-dimensional space scene. Come for a richer VR panoramic experience.
  • VR panoramic technology can be used to set a three-dimensional space scene for a house to be rented or a house to be sold. Any user can access through the network and watch the three-dimensional space scene of the corresponding house anytime and anywhere.
  • the present disclosure allows the user to leave his own footprint information for the house he is browsing, and the present disclosure can target the user’s own footprint information and other information about the house. The footprint information left by the user for the house is presented to the user together.
  • the footprint information 120 left by other users for the three-dimensional space scene of the two-bedroom and one-living house includes: "I like this group of sofas, it's great”, “This decorative partition is good”, “This sofa is good, high-end class”, “Combination and collocation are very diligent, praise and praise", "The design of the tea table is very unique ⁇ the longest copywriting is 20 characters” and the three-dimensional model 110 shown in the upper right corner of Figure 1.
  • the user who browses the three-dimensional space scene of the house is presented with the footprint information 120 left by other users with respect to the three-dimensional space scene of the house, which helps the user understand other users’ feelings about the house, thereby helping to deepen the user’s experience of the house. Cognition, which in turn helps to improve the user’s browsing experience of the house.
  • the user can also express his own feelings about the house, that is, leave his own footprint information in the three-dimensional space scene.
  • the user can set footstep information such as "this pillar makes the house look more distinctive" at the position of the pillar shown in FIG.
  • the footprint information set by the user can be instantly displayed in the three-dimensional space scene shown in Figure 1, that is, the user can see the footprint information left by himself during the process of viewing the three-dimensional space scene of the house, which is conducive to enhancing user participation feel.
  • all other footprint information set by the user for the house that does not belong to the view shown in FIG. 1 can be presented to the user in the form of a bullet screen 130, which is beneficial to improve the user's ability to browse three-dimensional space scenes in other locations of the house. interest.
  • the technology for realizing the interaction of three-dimensional space scenes provided by the present disclosure can also be applied to various other scenes. For example, when a user browses a three-dimensional space scene of a library, it can target a book or chair or coffee in the library. Set the corresponding footprint information.
  • the footprint information set by the user for the book may be the user's impression of the book or the number of pages currently read by the user.
  • the scenarios where the technology for realizing the interaction of three-dimensional space scenes provided by the present disclosure can be applied will not be illustrated one by one.
  • FIG. 2 is a flowchart of an embodiment of a method for realizing interaction of a three-dimensional space scene of the present disclosure.
  • the method 200 of the embodiment shown in FIG. 2 includes steps 210 to 240. Each step is described separately below.
  • step 210 in response to detecting the user operation of setting the footprint information in the three-dimensional space scene, determine the first pixel in the current view corresponding to the user's current perspective in the three-dimensional space scene.
  • a three-dimensional space scene may refer to a space scene with a three-dimensional sense that is presented to the user by using a preset panoramic image and a three-dimensional model.
  • the three-dimensional space scene may be a three-dimensional space scene set for a library, a three-dimensional space scene set for a house, a three-dimensional space scene set for a cafe, or a three-dimensional space scene set for a shopping mall.
  • the embodiment of the present disclosure when the user triggers the function of setting the footprint information in the three-dimensional space scene, it can be detected that the user needs to set the footprint information in the three-dimensional space scene. For example, when the user clicks a button for setting footstep information or a corresponding option on a menu, the embodiment of the present disclosure can detect that the user needs to set footstep information in a three-dimensional space scene. For another example, the user can use a preset shortcut to trigger the function of setting footprint information in the three-dimensional space scene.
  • the user's footprint information may be information that can indicate that the user has visited the three-dimensional space scene. The footprint information can be considered as the visit trace information of the user.
  • the current perspective of the user in the three-dimensional space scene may refer to the position and angle at which the user currently views the three-dimensional space scene.
  • the user's current perspective in the three-dimensional space scene usually changes with the user's operation. For example, the user can control his current perspective in the three-dimensional scene by performing operations such as dragging on the touch screen.
  • the user's current perspective in the three-dimensional space scene determines the content/area of the panorama that the user can currently see, that is, the user's current perspective in the three-dimensional space scene determines the current view.
  • the first pixel point is one pixel point in the current view.
  • the first pixel point can be obtained according to a preset default rule.
  • the first pixel may be a specific pixel in the current view, or it may be any pixel in the current view.
  • step 220 the three-dimensional model corresponding to the first pixel is determined.
  • a three-dimensional space scene is generally formed by a plurality of three-dimensional models.
  • the three-dimensional space scene may also be formed by a three-dimensional model.
  • a pixel point in the current view seen by the user may be a representation of a point in the three-dimensional model.
  • a pixel in the current view that the user sees may not be a representation of any point in the three-dimensional model. That is to say, under normal circumstances, any point in any three-dimensional model in the three-dimensional space scene can be presented in the panorama, and the points in the panoramic image may not be all three-dimensional models in the three-dimensional scene. In the point.
  • the present disclosure does not exclude the possibility that some points in the three-dimensional model in the three-dimensional space scene are not presented in the panoramic image.
  • the three-dimensional model where the point is located is the three-dimensional model corresponding to the first pixel.
  • the three-dimensional model corresponding to the first pixel may be: the current view that is similar to the first pixel and is used to present the points in the three-dimensional model.
  • the three-dimensional model corresponding to other pixels that is to say, when the first pixel is used to present the points in the non-three-dimensional model, and the first pixel is not updated, the three-dimensional model corresponding to other pixels in the current view can be used as the three-dimensional model corresponding to the first pixel. Model.
  • step 230 the position of the user's footprint information in the three-dimensional model is determined, where the footprint information is used to display when the three-dimensional space scene is browsed.
  • the positions of the first pixel or the other pixels in the three-dimensional model can be obtained. This location is the location of the user's footprint information.
  • all three-dimensional models in a three-dimensional space scene may be respectively provided with their own three-dimensional coordinate systems, or may have the same three-dimensional coordinate system.
  • the position of the user's footprint information in the three-dimensional model can be represented by (x, y, z). That is, the user's footprint information can be deep.
  • step 240 the user's footprint information is set at the location.
  • setting the user's footprint information at the location may include: setting a three-dimensional model identifier and three-dimensional coordinates for the user's footprint information, and storing the three-dimensional model identifier, three-dimensional coordinates, and user's footprint information. Correspondence.
  • the user's footprint information may be used to display to browsing users (such as all browsing users or some browsing users) of the three-dimensional space scene.
  • the browsing user of the three-dimensional space scene may include the user who sets the footprint information.
  • the three-dimensional model corresponding to the first pixel and the position of the footprint information in the three-dimensional model are obtained, so that the user can set
  • the footprint information of can be associated with the corresponding position of the corresponding 3D model.
  • the footprint information includes: at least one of text, picture, audio, video, and a three-dimensional model.
  • the text can be considered as a message in the form of characters (such as text, letters, numbers, or symbols, etc.).
  • a picture can be considered as a message in the form of an image (such as a photo or emoticon, etc.).
  • Audio can be thought of as a voice message (also called memo, etc.).
  • Video can be thought of as a message in the form of an image.
  • the three-dimensional model can be considered as a three-dimensional message.
  • the user's footprint information may be referred to as the user's message.
  • a piece of footprint information set by the user may include one or more of text, picture, audio, video, and three-dimensional model at the same time.
  • the user's footprint information By enabling the user's footprint information to include at least one of text, pictures, audio, video, and a three-dimensional model, it is beneficial to enrich the expression form of the user's footprint information, thereby helping to enrich the way for the user to interact with the three-dimensional space scene.
  • obtaining the first pixel point in the current view corresponding to the current perspective of the user in the three-dimensional space scene may be: obtaining the central pixel point of the current view corresponding to the current perspective of the user in the three-dimensional space scene , And use the central pixel as the first pixel.
  • the user’s current perspective in the 3D space scene triggers the function of setting footprint information in the 3D space scene by clicking a button or an option on the menu.
  • the center pixel can be considered as the default pixel set for the user's footprint information, and the user can change the default pixel by dragging and other methods.
  • the central pixel can be considered as a pixel in the central area of the current view.
  • the central area of the current view may include one pixel or multiple pixels.
  • obtaining the first pixel in the current view corresponding to the current perspective of the user in the three-dimensional space scene may be: setting the footprint information in the current view corresponding to the current perspective of the user in the three-dimensional space scene
  • the operation of the target position obtains the pixel point in the current view corresponding to the target position of the footprint information, and the pixel point is regarded as the first pixel point. That is, when the user performs the operation of setting the target position of the footprint information, the pixel point where the target position of the footprint information formed by the operation in the current view is located may be used as the first pixel point.
  • the operation of setting the target position of the footprint information may be an operation used to determine the starting target position of the footprint information, an operation used to determine the ending target position of the footprint information, or an operation used to determine the footprint information. Operation of the center target position.
  • the operation of setting the target location of the footprint information may specifically be a click operation or a scroll operation or drag operation based on a tool such as a mouse or a keyboard, and may also be a click operation or a drag operation based on a touch screen.
  • the present disclosure does not limit the specific operation of setting the target position of the footprint information.
  • the footprint information set by the user By determining the first pixel point according to the user's operation of setting the target position of the footprint information, it is beneficial to make the footprint information set by the user be located at the desired position of the user, thereby improving the flexibility of setting the footprint information and making the footprint information more stable. The location is more appropriate.
  • the user triggers the function of setting footprint information in the 3D space scene by clicking a button or an option on the menu.
  • the user can use the left mouse button to click, the keyboard up, down, left, and right keys to move the cursor or click the corresponding position on the touch screen to set the position of the desired footprint information in the current view.
  • the pixel at this position can be used as the first pixel.
  • the user triggers the function of setting footprint information in the 3D space scene by clicking a button or an option on the menu.
  • the pixel point is regarded as the first pixel point.
  • step 220 the implementation of determining the three-dimensional model corresponding to the first pixel (step 220) may be as shown in FIG. 3. As shown in FIG. 3, step 220 further includes steps 310 to 340.
  • step 310 the central pixel point of the front view is determined as the first pixel point.
  • the central pixel may be considered as the default pixel set for the user's footprint information.
  • the current view is an image of (2n+1) ⁇ (2m+1) (where n and m are both integers greater than 1), then the pixels in the current view (n+1 , M+1) as the central pixel.
  • the pixel point (n, m) and pixel point (n+1) in the current view can be , M), pixel (n, m+1), and pixel (n+1, m+1) are used as the central area of the current view, so that any pixel in the central area can be used as the central pixel.
  • step 320 it is determined whether a three-dimensional model is set for the first pixel. If a three-dimensional model is set for the first pixel point, go to step 330. If the three-dimensional model is not set for the first pixel point, go to step 340.
  • step 330 in response to the determination that the three-dimensional model is set for the first pixel, the three-dimensional model set for the first pixel is used as the three-dimensional model corresponding to the first pixel.
  • step 340 in response to the determination that the three-dimensional model is not set for the first pixel, the three-dimensional model set for other pixels in the current view is used as the three-dimensional model corresponding to the first pixel.
  • pixels in the current view are pixels in the current view where the three-dimensional model is set.
  • the pixels where the three-dimensional model is set can be found according to preset rules.
  • the other pixels found may be the pixels closest to the first pixel in a certain direction (such as the left direction, the right direction, the upper direction, or the lower direction).
  • the first pixel point can be used as a starting point, and according to a preset inspection rule, the pixel point in the current view corresponding to the current perspective in the three-dimensional space scene can be checked. If it is determined that the pixel point with the three-dimensional model is checked , The three-dimensional model corresponding to the first pixel is obtained, and the inspection process is stopped. For example, you can use the first pixel as a starting point to check the pixels in the current view to the left, and determine whether a three-dimensional model is set for the currently checked pixel.
  • the inspection process is stopped, and the three-dimensional model obtained by the current inspection is used as the three-dimensional model corresponding to the first pixel.
  • the first pixel point may be updated by using the detected pixel point provided with the three-dimensional model.
  • the first pixel may not be updated.
  • step 220 may be as shown in FIG. 4.
  • step 220 may include step 410 to step 450.
  • step 410 in response to the user's operation of setting the target position of the footprint information in the current view, a pixel point in the current view corresponding to the target position of the footprint information is determined as the first pixel point.
  • the user may be allowed to set the specific location of the footprint information (that is, the target location of the footprint information) in the current view.
  • the user can set the target position of the footprint information in the current view by tapping, sliding, dragging and other operations on the touch screen.
  • the target position of the footprint information may be the upper left vertex, the lower left vertex, the upper right vertex, or the lower right vertex of the text box.
  • the target location of the footprint information may be the upper left vertex, the lower left vertex, the upper right vertex, or the lower right vertex of the picture.
  • the target location of the footprint information may be a pixel point in the current view, and the pixel point is the first pixel point.
  • step 420 it is determined whether a three-dimensional model is set for the first pixel. If a three-dimensional model is set for the first pixel point, go to step 430. If a three-dimensional model is not set for the first pixel point, go to step 440.
  • step 430 in response to the determination that the three-dimensional model is set for the first pixel, the three-dimensional model set for the first pixel is used as the three-dimensional model corresponding to the first pixel.
  • step 440 in response to the determination that the three-dimensional model is not set for the first pixel point, prompt information for updating the target position of the footprint information is output.
  • the prompt information is used to prompt the user to update the target location of the footprint information currently set. That is, the prompt information is used to prompt the user that the current target location of the footprint information cannot be set to the footprint information, and the user should reset the target location of the footprint information.
  • the prompt information can be output in the form of text, audio, or graphics. After outputting the prompt information, wait for the user's subsequent operations. If the user triggers the function of canceling the setting of the footprint information at this time, the process shown in Figure 4 ends.
  • step 450 in response to the determination that the pixel in the current view corresponding to the target position of the updated footprint information is set with a three-dimensional model, the pixel set with the three-dimensional model is taken as the first pixel. Then, the flow returns to step 420.
  • the target position of the footprint information obtained again may also be a pixel in the current view, and this pixel is the first pixel. That is, the first pixel point obtained last time is updated by the similarity of the target position of the currently obtained footprint information.
  • the first pixel when the first pixel is provided with a three-dimensional model, since the first pixel in the current view has a mapping relationship with the point in the three-dimensional model, the first pixel can be obtained based on the mapping relationship.
  • the point corresponding to the point in the three-dimensional model the position of the point is the position of the first pixel point in the three-dimensional model.
  • the position of the first pixel in the three-dimensional model can be directly used as the position of the user's footprint information in the three-dimensional model, which is beneficial to quickly and accurately obtain the position of the user's footprint information in the three-dimensional model.
  • the browsing user in the process of viewing the three-dimensional space scene by the browsing user, the browsing user may be presented with at least one user's footprint information left in the three-dimensional space scene.
  • An example is shown in Figure 5.
  • step 510 for any browsing user browsing the three-dimensional space scene, the footprint area corresponding to the current perspective of the browsing user in the three-dimensional space scene is determined.
  • the browsing user includes a user who sets his footprint information in the three-dimensional space scene.
  • the footprint area can be considered as an area set for the footprint information that needs to be displayed.
  • the footprint area can be a footprint area based on the current view, or a footprint area based on a three-dimensional model.
  • the size of the footprint area can be preset.
  • the shape of the footprint area can be rectangle, circle, triangle, etc.
  • an implementation manner of determining the footprint area may be: First, obtain the current view corresponding to the current viewing angle of the browsing user in the three-dimensional space scene. The center pixel of the view, and then the center pixel is the center of the circle, and the predetermined length (such as 1.5 meters in a three-dimensional scene, etc., and 1.5 meters can be converted to the length in the current view) is used as the radius to determine the current view Footprint area. Since at least some of the pixels in the footprint area in the current view have a mapping relationship with points in the three-dimensional model, the footprint information that currently needs to be displayed can be easily obtained by using the footprint area in the current view. In addition, the footprint area in the current view can be regarded as a circle, that is, the footprint area in the current view does not have depth information.
  • an implementation manner for determining the footprint area may be: First, obtain the current view corresponding to the current perspective of the browsing user in the three-dimensional space scene. View the center pixel of the view, and determine whether the center pixel is set with a three-dimensional model. If the center pixel is set with a three-dimensional model, determine the position of the center pixel in the three-dimensional model, and then use that position as the center of the circle, Use a predetermined length (such as 1.5 meters in a three-dimensional space scene, etc.) as a radius to determine the footprint area in the three-dimensional model.
  • the footprint area may be completely in one 3D model, or it may span multiple 3D models.
  • the footprint area in the three-dimensional model can be considered as a cylinder, that is, the footprint area in the three-dimensional model has depth information.
  • step 520 the footprint information belonging to the footprint area in the three-dimensional model is determined.
  • the embodiment of the present disclosure can check whether each pixel point in the footprint area has a mapping relationship with a point in the three-dimensional model. If there is a mapping relationship, then it is determined whether the points in the three-dimensional model that have a mapping relationship with the pixel points are provided with footprint information. If the footprint information is set, the footprint information can be regarded as the footprint information belonging to the footprint area.
  • the embodiments of the present disclosure can check whether each point in the footprint area is provided with footprint information. If the footprint information is set, the footprint information can be regarded as the footprint information belonging to the footprint area.
  • step 530 in the current view corresponding to the current perspective of the browsing user in the three-dimensional space scene, the footprint information belonging to the footprint area is displayed.
  • the location of each footprint information belonging to the footprint area in the current view can be determined according to the location of each footprint information, so that each footprint information can be displayed according to the location of each footprint information in the current view.
  • the process of displaying footprint information it is possible to avoid overlapping display of different footprint information in the current view.
  • the obtained multiple footprint information may have different positions, or may have the same position (that is, the position of the footprint information conflicts).
  • each footprint information may be displayed in the current view directly according to the image positions of the multiple footprint information in the current view.
  • the displayed footprint information can be allowed to partially overlap, and the location control can also be used to make the footprint information not overlap each other.
  • different image positions may be assigned to different footprint information in the current view, and the image positions described above may be displayed in the current view according to the assigned image positions. Having different footprint information at the same position helps to avoid overlapping display of different footprint information in the current view.
  • all the footprint information belonging to the footprint area can be displayed, or part of the footprint information belonging to the footprint area can be displayed.
  • part of the footprint information can be selected from it according to a predetermined rule, and the selected part of the footprint information can be displayed in the current view.
  • a predetermined number of footprint information can be randomly selected from all the footprint information belonging to the footprint area, and part of the randomly selected footprint information can be displayed in the current view.
  • the form of a bullet screen may be used to display the footprint information outside the current view for the browsing user. For example, you can first determine all the footprint information in the three-dimensional model that does not belong to the current view, and display all the above-mentioned footprint information in the current view corresponding to the current perspective of the browsing user in the three-dimensional scene in the form of a bullet screen. Part of the footprint information.
  • the form of a bullet screen may be used to display the footprint information outside the footprint area for the browsing user. For example, you can first determine all the footprint information in the three-dimensional model that does not belong to the footprint area, and display all the above-mentioned footprint information in the current view corresponding to the current perspective of the browsing user in the three-dimensional scene in the form of a bullet screen. Part of the footprint information.
  • FIG. 6 is a schematic structural diagram of an embodiment of an apparatus for realizing interaction in a three-dimensional space scene of the present disclosure.
  • the device of this embodiment can be used to implement the foregoing method embodiments of the present disclosure.
  • the device of this embodiment includes: a pixel point acquiring module 600, a three-dimensional model determining module 601, a position determining module 602, and a footprint information setting module 603.
  • the device may further include: a footprint area determination module 604, a footprint information determination module 605, a footprint information display module 606, and a bullet screen display module 607.
  • the pixel obtaining module 600 is configured to determine the first pixel in the current view corresponding to the current perspective of the user in the three-dimensional space scene in response to detecting the user operation of setting the footprint information in the three-dimensional space scene.
  • the footprint information may include: at least one of text, picture, audio, video, and a three-dimensional model.
  • the pixel point acquiring module 600 may include: a first sub-module 6001.
  • the first sub-module 6001 is used to determine the center pixel of the current view as the first pixel.
  • the pixel point obtaining module 600 may include: a fifth sub-module 6002.
  • the fifth sub-module 6002 is configured to determine the pixel points in the current view corresponding to the target position of the footprint information in response to the user's operation of setting the target position of the footprint information in the current view corresponding to the current perspective in the three-dimensional space scene.
  • the fifth sub-module 6002 can use the pixel as the first pixel.
  • the three-dimensional model determining module 601 is used to determine the three-dimensional model corresponding to the first pixel obtained by the pixel obtaining module 600.
  • the determining three-dimensional model module 601 may include: the second sub-module 6011, the third sub-module 6012, and the fourth sub-module 6013.
  • the second sub-module 6011 is used to determine whether a three-dimensional model is set for the first pixel.
  • the third sub-module 6012 is configured to, if the determination result of the second sub-module 6011 is that a three-dimensional model is set for the first pixel, use the three-dimensional model set for the first pixel as the three-dimensional model corresponding to the first pixel.
  • the fourth sub-module 6013 is configured to, if the judgment result of the second sub-module 6011 is that no three-dimensional model is set for the first pixel, use the three-dimensional model set for other pixels in the current view as the three-dimensional model corresponding to the first pixel. Model. For example, if the judgment result of the second sub-module 6011 is that a three-dimensional model is not set for the first pixel, the fourth sub-module 6013 can take the first pixel as a starting point and perform a check on the three-dimensional scene according to the preset inspection rules. Check other pixels in the current view corresponding to the current angle of view. If a pixel with a three-dimensional model is detected, the first pixel is updated to a pixel with a three-dimensional model, the three-dimensional model corresponding to the first pixel is obtained, and this inspection is stopped.
  • the determining three-dimensional model module 601 may include: a sixth sub-module 6014, a seventh sub-module 6015, and an eighth sub-module 6016.
  • the sixth sub-module 6014 is used to determine whether a three-dimensional model is set for the first pixel. If the determination result of the sixth sub-module 6014 is that a three-dimensional model is set for the first pixel, the seventh sub-module 6015 uses the three-dimensional model set for the first pixel as the three-dimensional model corresponding to the first pixel.
  • the eighth sub-module 6016 may output prompt information for updating the target position of the footprint information, and the sixth sub-module 6014 determines When the pixel in the current view corresponding to the target location of the footprint information is set with a three-dimensional model, the pixel with the three-dimensional model is used as the first pixel.
  • the eighth sub-module 6016 obtains the three-dimensional model corresponding to the first pixel.
  • the position determining module 602 is used to determine the position of the user's footprint information in the three-dimensional model determined by the three-dimensional model determining module 601. For example, the position determining module 602 may obtain the position of the first pixel in the three-dimensional model, and the position determining module 602 may use the position of the first pixel in the three-dimensional model as the position of the user's footprint information in the three-dimensional model.
  • the setting footprint information module 603 is used for setting the user's footprint information at the location determined by the location determining module 602.
  • the user's footprint information set by the setting footprint information module 603 is used to be displayed to users who browse the three-dimensional space scene.
  • the footprint area determination module 604 is used for determining the footprint area corresponding to the current perspective of the browsing user in the three-dimensional space scene for any browsing user who browses the three-dimensional space scene. For example, the module 604 for determining the footprint area may first determine the center pixel of the current view corresponding to the current perspective of the browsing user in the three-dimensional space scene, and then the module 604 for determining the footprint area 604 takes the center pixel as the center of the circle and the predetermined length as The radius determines the footprint area in the current view.
  • the footprint information determining module 605 is used to determine the footprint information belonging to the footprint area determined by the footprint area determining module 604 in the three-dimensional model.
  • the footprint information display module 606 is configured to display the footprint information that belongs to the footprint area determined by the determination footprint information module 605 in the current view corresponding to the current perspective of the browsing user in the three-dimensional space scene.
  • the display footprint information module 606 may respectively locate the multiple footprint information in the current view according to the image positions of the multiple footprint information in the current view. Display the multiple footprint information.
  • the footprint information display module 606 may assign different image positions for different footprint information in the current view, and according to the assigned image Position, display different footstep information in the current view.
  • the barrage display module 607 is used to determine at least one piece of footprint information in the three-dimensional model that does not belong to the footprint area/current view.
  • the barrage display module 607 displays the at least one footprint information in the current view corresponding to the current perspective of the browsing user in the three-dimensional space scene in the form of a barrage.
  • FIG. 7 shows a process 700 of an embodiment of the first three-dimensional model-based interaction method according to the present disclosure.
  • the three-dimensional model-based interaction method is applied to a first user terminal, and the first user terminal is presented with a user interface, and the three-dimensional model-based interaction method includes:
  • Step 710 In response to detecting the user's target interaction operation on the user interface, send an interaction request for the target interaction operation to the server that provides page data for the user interface, where the user interface is used to present the three-dimensional model, the three-dimensional model and the second user The user account logged in by the terminal establishes an association relationship.
  • the user can use the first user terminal to interact with the server through the network.
  • the first user terminal may be various electronic devices, including but not limited to smart phones, tablet computers, laptop portable computers, desktop computers, and so on.
  • the first user terminal may be installed with various client applications, such as real estate transaction software.
  • the aforementioned user interface may be a page in an application installed by the first user terminal.
  • the user can interact with the server through the user interface, thereby realizing interaction with other user terminals (for example, the second user terminal).
  • the first user terminal may send an interaction request for the target interaction operation to a server that provides page data for the user interface.
  • the aforementioned user interface is used to present a three-dimensional model.
  • the three-dimensional model establishes an association relationship with the user account logged in by the second user terminal in advance.
  • the aforementioned target interaction operation may be various operations for instructing the first user terminal to request interaction (information interaction) with the second user terminal.
  • the target interaction operation may indicate video communication with the second user terminal.
  • the foregoing interaction request may be used to indicate a user request of the first user terminal to interact with the second user terminal.
  • the foregoing interaction request may be used to instruct the user of the first user terminal to request video communication with the second user terminal.
  • the user interface of the first user terminal may present the above-mentioned three-dimensional model, or may not present the three-dimensional model.
  • each three-dimensional model it can be associated with a user account in advance. Therefore, for a specific three-dimensional model, the user account that is associated with the three-dimensional model can be determined to determine the user terminal logging in to the user account, and then the user terminal used to interact with the first user terminal is determined (Ie the second user terminal).
  • the above-mentioned three-dimensional model may be a three-dimensional model of any object.
  • the three-dimensional model may be a three-dimensional model inside a cell, or a three-dimensional model of a house interior.
  • Step 720 Receive the streaming video obtained by the server from the second user terminal.
  • the above-mentioned first user terminal may receive the streaming video obtained by the server from the second user terminal.
  • the aforementioned interaction confirmation information may be used to instruct the user of the second user terminal to confirm (agree) to perform the interaction indicated by the aforementioned interaction request with the first user terminal.
  • the foregoing interactive confirmation information may be used to instruct the user of the second user terminal to confirm (agree) to conduct video communication with the first user terminal.
  • the aforementioned streaming video may include images and/or voice.
  • the image acquisition device and/or the voice acquisition device of the second user terminal can be used to acquire the aforementioned streaming video.
  • the server may use streaming media technology to continuously send the images and/or voice (ie streaming media video) collected by the second user terminal to the first user terminal.
  • streaming media technology refers to a media format that uses streaming technology to be continuously played in real time on the network.
  • Streaming media technology is also called streaming media technology.
  • the second user terminal may send the continuous image and sound information collected by it to the server after compression processing.
  • the server transmits each compressed package to the first user terminal sequentially or in real time, so that users who use the first user terminal can watch and listen while downloading.
  • the server may send the streaming video collected by the second user terminal to the first user terminal, and may also perform image processing (such as beauty) and voice processing (such as denoising) on the streaming video collected by the second user terminal. ), after operations such as transcoding, recording, and pornography, send the processed streaming video to the first user terminal.
  • image processing such as beauty
  • voice processing such as denoising
  • the first user terminal may perform step 720 again.
  • the first user terminal in the case of the interaction confirmation information sent by the second user terminal in response to the interaction request, can present the streaming video through subsequent steps; however, the second user terminal does not send the above
  • the first user terminal does not present the streaming video. Therefore, the streaming video and the three-dimensional model can be presented on the user interface of the first user terminal only after obtaining the permission of the user of the second user terminal (for example, connecting to the video call initiated by the first user terminal). This helps to improve the privacy protection of the user of the second user terminal, and provides preparation time for the user of the second user terminal to present the streaming media video to the user of the first user terminal.
  • the first user terminal may also directly execute the foregoing step 720 (without the interaction confirmation information sent by the second user terminal in response to the interaction request).
  • the user of the second user terminal may be in a state of shooting a streaming video (for example, a live broadcast) to users of other user terminals.
  • a streaming video for example, a live broadcast
  • the first user terminal can receive the streaming video obtained by the server from the second user terminal at any time, thereby improving the real-time performance of the streaming video presentation.
  • the first user terminal may adopt the following steps to receive the streaming video obtained by the server from the second user terminal:
  • the current network speed value of the first user terminal is sent to the server.
  • the streaming media video obtained and sent by the server from the second user terminal is received, and the streaming media video has a resolution matching the current network speed value.
  • the resolution can be positively correlated with the network speed value.
  • the resolution of the streaming video received by the first user terminal can be reduced when the network is poor by receiving the streaming video sent by the server whose resolution matches the current network speed value and obtained by the second user terminal. Rate to improve the real-time performance of streaming video transmission.
  • Step 730 Present the streaming video and the three-dimensional model on the user interface.
  • the first user terminal may present the streaming video and the three-dimensional model on the same screen on the user interface.
  • the above-mentioned user interface of the first user terminal may be divided into two parts, and the above-mentioned two parts may respectively present a streaming video and a three-dimensional model.
  • the three-dimensional model can also be used as the background of the aforementioned user interface, and the streaming video is presented in a part of the page area of the user interface.
  • FIGS. 8A-8C are schematic diagrams of application scenarios for the embodiment of FIG. 7.
  • the first user terminal may provide the user interface with The server of the page data sends an interaction request for the target interaction operation 810.
  • the user interface presents a three-dimensional model of the house of XX home.
  • the three-dimensional model has a pre-established association relationship with the user account logged in by the second user terminal.
  • the first user terminal presents a streaming video 830 and a three-dimensional model on the user interface.
  • the interaction method based on the three-dimensional model provided by the above-mentioned embodiments of the present disclosure can send an interaction request for the target interaction operation to a server that provides page data for the user interface when the user's target interaction operation for the user interface is detected.
  • the user interface is used for presenting a three-dimensional model, and the three-dimensional model establishes an association relationship with the user account logged in by the second user terminal in advance.
  • the streaming video obtained by the server from the second user terminal is received.
  • the streaming video and 3D model are presented on the user interface.
  • the first user terminal may also perform the following steps:
  • the model adjustment information sent by the server is received, where the model adjustment information indicates an adjustment operation of the user who uses the second user terminal on the three-dimensional model presented on the second user terminal.
  • the adjustment operation includes at least one of the following: zoom, rotate, move, and switch viewpoints.
  • the user can perform at least one operation of zooming, rotating, moving, and switching viewpoints on the three-dimensional model.
  • the same adjustment operation is performed on the three-dimensional model presented on the user interface.
  • the operations performed by the user of the second user terminal on the three-dimensional model can be synchronized to the first user terminal. Therefore, when the streaming video collected by the second user terminal is related to the three-dimensional model (for example, the user of the second user terminal explains or introduces the three-dimensional model), it is convenient for the user of the first user terminal to refer to the second user The same three-dimensional model presented by the terminal acquires the information in the streaming video, thereby improving the pertinence of information acquisition.
  • the first user terminal may also perform the following steps:
  • the feedback information may include but is not limited to at least one of the following: likes, ratings, comments, and so on.
  • the feedback information may be used to characterize the evaluation of the user of the first user terminal on the streaming video of the user of the second user terminal.
  • the feedback information is sent to the server, where the server is used to establish an association relationship between the feedback information and the user account.
  • the server is used to establish an association relationship between the feedback information and the user account.
  • an associative storage method can be used to establish an association relationship between the feedback information and the user account.
  • establishing an association relationship between the feedback information and the user account can reflect the user's satisfaction with the object indicated by the three-dimensional model and the user of the second user terminal by the user of the first user terminal, and thus can be more targeted for the first user The terminal pushes information.
  • FIG. 9 is a flow 900 of another embodiment of the first three-dimensional model-based interaction method of the present disclosure.
  • the three-dimensional model-based interaction method is applied to a first user terminal, and the first user terminal is presented with a user interface, and the method includes:
  • Step 910 In response to detecting the user's target interaction operation on the user interface, send an interaction request for the target interaction operation to a server that provides page data for the user interface.
  • Step 920 Receive the streaming video obtained by the server from the second user terminal.
  • Step 930 Present the streaming video and the three-dimensional model on the user interface.
  • step 910 to step 930 are basically the same as step 710 to step 730 in the embodiment corresponding to FIG. 7, and will not be repeated here.
  • Step 940 In response to the current network speed value of the first user terminal being less than or equal to the preset network speed threshold, adjust the target user image based on each frame of voice in the streaming video to generate a new video different from the streaming video.
  • the first user terminal may adjust the target user image based on each frame of voice in the streaming video, Generate a new video.
  • the new video characterizes the actions of the user indicated by the target user's image to perform each frame of voice instructions.
  • the user indicated by the target user image may be a user using the second user terminal.
  • the new video may be a streaming video that is sent in segments and instantly transmitted based on the network, or it may be a video that is generated locally without being based on the network.
  • the first user terminal may generate a new video in the following manner: For each frame of voice in the streaming video, input the frame of voice into a predetermined image frame generation model to obtain a target user that matches the frame of voice The image of the user indicated by the image. Thereby, the obtained frames of images that match each frame of voice in the streaming video and the frames of voice are merged to obtain a new video. The user's action in the user's image indicated by the target user image that matches the voice matches the voice.
  • the mouth shape of the user in the user's image indicated by the target user image matching the audio may be the voice "ah” Lip shape
  • the action can be an action in a state of fright.
  • the aforementioned image frame generation model may be a recurrent neural network model or a convolutional neural network model obtained by training using a machine learning algorithm based on training samples including voice frames, target user images, and image frames matching the voice frames.
  • An image frame generation model can be trained for each user, and the target user image in each training sample used to train the user’s image frame generation model can be the same.
  • For each voice frame of the user it is determined that the voice frame corresponds to the voice frame. Matched image frames, and then obtain a training sample set used to train the image frame generation model of the user.
  • the image frame generation model may also be a two-dimensional table or database that stores the voice frame, the target user image, and the image frame matching the voice frame in association with each other.
  • each record of the database may include voice frames, target user images and matching voice frames Image frame.
  • the target user image in each record can be the same.
  • Frame database that is, image frame generation model.
  • the first user terminal may also determine the target user image by any of the following methods:
  • the target user image is generated.
  • the image whose ratio is greater than the preset threshold is regarded as the target user image.
  • the user can upload an image through the user account he uses as the target user image; or after logging in the account he uses, select an image from a predetermined image set as the target user image.
  • the above-mentioned optional implementation methods can automatically generate a target user image from the images in the streaming video, or the user manually sets the target user image, so that based on multiple target user image determination methods, the new video
  • the generation method is more diversified.
  • step 950 the new video is used to replace the streaming video for presentation.
  • the first user terminal may use a new video to replace the streaming video for presentation.
  • the streaming video can be hidden (that is, no longer presented).
  • the first The user terminal can locally generate a new video to replace the streaming video presentation. Therefore, the first user terminal only needs to continuously obtain voice from the server, but does not need to continuously obtain video, thereby reducing the occupation of network resources. In a case where the current network speed value of the first user terminal is relatively small, the real-time performance of the video presentation of the first user terminal can be improved.
  • the first user terminal may also send a camera shutdown confirmation to the server. information.
  • the camera closing confirmation information is used to determine whether the second user terminal closes the camera.
  • the server may send to the second user terminal information for determining whether the second user terminal turns off the camera. Therefore, the user of the second user terminal can reduce the occupation of network resources by the second user terminal by turning off the camera.
  • FIG. 10 is a flowchart of another embodiment of the first three-dimensional model-based interaction method of the present disclosure.
  • the interaction method based on the three-dimensional model is applied to a first user terminal, and the first user terminal presents a user interface.
  • the process 1000 of the interaction method based on the three-dimensional model includes:
  • Step 1010 In response to detecting the user's target interaction operation on the user interface, send an interaction request for the target interaction operation to a server that provides page data for the user interface.
  • the user interface is used for presenting a three-dimensional model, and the three-dimensional model establishes an association relationship with the user account logged in by the second user terminal in advance.
  • Step 1020 Receive the streaming video obtained by the server from the second user terminal.
  • Step 1030 Present the streaming video and the three-dimensional model on the user interface.
  • step 1010 to step 1030 are basically the same as step 710 to step 730 in the embodiment corresponding to FIG. 7, and will not be repeated here.
  • the three-dimensional model includes three-dimensional sub-models of multiple sub-space scenes, and the sub-space scenes in the multiple sub-space scenes correspond to keywords in a predetermined keyword set.
  • Step 1040 Perform voice recognition on the voice in the streaming video to obtain a voice recognition result.
  • the first user terminal may perform voice recognition on the voice in the streaming video to obtain the voice recognition result.
  • the voice recognition result can represent the text corresponding to the voice in the streaming video.
  • Step 1050 In response to the determination that the voice recognition result includes keywords in the keyword set, present on the user interface a three-dimensional sub-model of the corresponding sub-space scene among the multiple sub-space scenes corresponding to the keywords included in the voice recognition result .
  • the first user terminal may display the subspace scene corresponding to the keywords contained in the voice recognition result on the aforementioned user interface.
  • the above-mentioned three-dimensional model is a three-dimensional model inside a house.
  • the house includes a bedroom, a living room, a kitchen, and a bathroom, with a total of four sub-space scenes. That is, the above-mentioned three-dimensional model includes a three-dimensional sub-model of a bedroom, a three-dimensional sub-model of a living room, a three-dimensional sub-model of a kitchen, and a three-dimensional sub-model of a bathroom.
  • the keyword set includes bedroom, living room, kitchen, bathroom.
  • the keyword corresponding to the subspace scene bedroom can be "bedroom”; the keyword corresponding to the subspace scene kitchen can be “kitchen”; the keyword corresponding to the subspace scene living room can be “living room” ; The keyword corresponding to the bathroom in the subspace scene can be "toilet”. Further, as an example, if the voice recognition result includes the keyword "bedroom", then the first user terminal may present a three-dimensional sub-model of the bedroom on the aforementioned user interface.
  • the embodiment of the present application may also include the same or similar features and effects as the embodiment corresponding to FIG. 7 and/or FIG. 9, and details are not described herein again.
  • the viewpoint switching of the three-dimensional model can be realized by voice, thereby presenting the subspace scene corresponding to the keywords contained in the voice recognition result The three-dimensional sub-model.
  • the convenience of browsing the three-dimensional model is improved, and the matching between the presented three-dimensional model and the voice acquired by the second user terminal is improved.
  • FIG. 11 shows a process 1100 of an embodiment of the second three-dimensional model-based interaction method according to the present disclosure.
  • the three-dimensional model-based interaction method is applied to a second user terminal, and the user account logged in by the second user terminal establishes an association relationship with the three-dimensional model in advance.
  • the interactive method based on the 3D model includes:
  • Step 1110 In response to receiving the interactive request sent by the server, obtain the streaming video.
  • the user can use the second user terminal to interact with the server and the first user terminal through the network.
  • the second user terminal may be various electronic devices, including but not limited to smart phones, tablet computers, laptop computers, desktop computers, and so on.
  • the second user terminal may be installed with various client applications, such as real estate transaction software.
  • the streaming video upon receiving the interaction request sent by the server, the streaming video is acquired.
  • the interaction request indicates that the first user terminal detects the user's target interaction operation on the user interface presented by the first user terminal.
  • the aforementioned interaction request may be used to instruct the user of the first user terminal to request video communication with the second user terminal.
  • the user interface is used to present the three-dimensional model.
  • Streaming videos can contain images and/or voice.
  • the image acquisition device and/or the voice acquisition device of the second user terminal can be used to acquire the aforementioned streaming video.
  • the first user terminal may send an interaction request for the target interaction operation to a server that provides page data for the user interface.
  • the user interface is used to present the three-dimensional model.
  • the three-dimensional model establishes an association relationship with the user account logged in by the second user terminal in advance.
  • the aforementioned target interaction operation may be various operations for instructing the first user terminal to request interaction (information interaction) with the second user terminal.
  • the target interaction operation may indicate video communication with the second user terminal.
  • the user interface of the first user terminal may present the above-mentioned three-dimensional model, or may not present the three-dimensional model.
  • Step 1120 Send the streaming video to the server.
  • the second user terminal may send the streaming video to the server.
  • the server is used to send the streaming video to the first user terminal, so that the first user terminal presents the streaming video and the three-dimensional model on the user interface.
  • the server can use streaming media technology to continuously send the images and/or voice (that is, streaming video) collected by the second user terminal to the first user terminal.
  • Streaming media technology refers to a media format that uses streaming technology to continuously play on the network in real time.
  • the second user terminal may send the continuous image and sound information collected by it to the server after compression processing.
  • the server transmits each compressed package to the first user terminal sequentially or in real time, so that users who use the first user terminal can watch and listen while downloading.
  • the server may send the streaming video collected by the second user terminal to the first user terminal, and may also perform image processing (such as beauty) and voice processing (such as denoising) on the streaming video collected by the second user terminal. ), after operations such as transcoding, recording, and pornography, send the processed streaming video to the first user terminal.
  • image processing such as beauty
  • voice processing such as denoising
  • the second three-dimensional model-based interaction method provided by the foregoing embodiment of the present disclosure is applied to a second user terminal, and the user account logged in by the second user terminal establishes an association relationship with the three-dimensional model in advance.
  • the second user terminal may determine whether the user's confirmation operation for the interaction request is detected in the case of receiving the interaction request sent by the server.
  • the interaction request indicates that the first user terminal detects the user's target interaction operation on the user interface presented by the first user terminal, and the user interface is used to present a three-dimensional model. Afterwards, if the confirmation operation is detected, the streaming video is obtained.
  • the streaming video is sent to the server, where the server is used to send the streaming video to the first user terminal, so that the first user terminal presents the streaming video and the three-dimensional model on the user interface.
  • the server is used to send the streaming video to the first user terminal, so that the first user terminal presents the streaming video and the three-dimensional model on the user interface.
  • the streaming media video by presenting the streaming media video and the three-dimensional model on the same page of the terminal device, it is helpful to use the streaming media video to present information related to the three-dimensional model to the user, thereby increasing the diversity of interaction modes.
  • users can browse the three-dimensional model more calmly, improve the user's browsing time, and help meet the users' more diversified interactive needs.
  • the foregoing step 1110 may include the following steps:
  • the confirmation operation indicates that the user of the second user terminal confirms (agrees) to interact with the first user terminal (for example, video communication).
  • the first user terminal may present the streaming video; and the second user terminal does not send the interaction confirmation information.
  • the first user terminal does not present the streaming video. Therefore, the streaming video and the three-dimensional model can be presented on the user interface of the first user terminal only after the permission of the user of the second user terminal is obtained (for example, the video call initiated by the first user terminal is connected). This helps to improve the privacy protection of the user of the second user terminal, and provides preparation time for the user of the second user terminal to present the streaming media video to the user of the first user terminal.
  • the second user terminal may also directly obtain the streaming video, and send the streaming video to the first user terminal through the server, without the need for a second user terminal. 2. Interaction confirmation information sent by the user of the user terminal in response to the interaction request.
  • the user of the second user terminal may be in a state of shooting a streaming video (for example, a live broadcast) to users of other user terminals.
  • a streaming video for example, a live broadcast
  • the first user terminal can receive the streaming video obtained by the server from the second user terminal at any time, thereby improving the real-time performance of the streaming video presentation.
  • the second user terminal may receive the camera shutdown confirmation message from the server, and display The camera is turned off confirmation message.
  • the camera close confirmation information is used to determine whether the second user terminal closes the camera.
  • the server may send to the second user terminal information for determining whether the second user terminal turns off the camera. Therefore, the user of the second user terminal can reduce the occupation of network resources by the second user terminal by turning off the camera.
  • the second user terminal may send model adjustment information indicating the adjustment operation to the server , So that the server controls the first user terminal to perform the same adjustment operation on the three-dimensional model presented on the user interface according to the adjustment operation indicated by the model adjustment information.
  • the adjustment operation includes at least one of the following: zoom, rotate, move, and switch viewpoints.
  • the user can perform at least one operation of zooming, rotating, moving, and switching viewpoints on the three-dimensional model.
  • the operations performed by the user of the second user terminal on the three-dimensional model can be synchronized to the first user terminal. Therefore, when the streaming video collected by the second user terminal is related to the three-dimensional model (for example, the user of the second user terminal explains or introduces the three-dimensional model), it is convenient for the user of the first user terminal to refer to the second user The same three-dimensional model presented by the terminal acquires the information in the streaming video, thereby improving the pertinence of information acquisition.
  • the second user terminal may follow the adjustment operation indicated by the model adjustment information , Perform the same adjustment operation on the three-dimensional model presented by the second user terminal.
  • the adjustment operation includes at least one of the following: zoom, rotate, move, and switch viewpoints.
  • the user can perform at least one operation of zooming, rotating, moving, and switching viewpoints on the three-dimensional model.
  • the operations performed by the user of the first user terminal on the three-dimensional model can be synchronized to the second user terminal.
  • the second user terminal may perform the same method as the feedback information. Matching operation.
  • the feedback information may include but is not limited to at least one of the following: likes, ratings, comments, and so on.
  • the feedback information may be used to characterize the evaluation of the user of the first user terminal on the streaming video of the user of the second user terminal.
  • the second user terminal may present an operation that matches the feedback information, for example, “XX gave you a like!” .
  • FIG. 12 is a flow 1200 of another embodiment of the second three-dimensional model-based interaction method of the present disclosure.
  • the three-dimensional model-based interaction method is applied to a first user terminal, and the first user terminal presents a user Interface, the method includes:
  • Step 1210 In response to receiving the interactive request sent by the server, obtain the streaming video.
  • Step 1220 Send the streaming video to the server.
  • step 1210 to step 1220 are basically the same as step 1110 to step 1120 in the embodiment corresponding to FIG. 11, and will not be repeated here.
  • the three-dimensional model includes three-dimensional sub-models of multiple sub-space scenes, and the sub-space scenes in the multiple sub-space scenes correspond to keywords in a predetermined keyword set.
  • Step 1230 Perform voice recognition on the voice acquired by the first user terminal to obtain a voice recognition result.
  • the second user terminal may perform voice recognition on the voice acquired by the first user terminal to obtain a voice recognition result.
  • the voice recognition result can represent the text corresponding to the voice in the streaming video.
  • Step 1240 in response to the determination that the voice recognition result contains keywords in the keyword set, present on the user interface a three-dimensional sub-model of the corresponding sub-space scene among the multiple sub-space scenes corresponding to the keywords contained in the voice recognition result .
  • the second user terminal may present on the user interface a three-dimensional subspace scene corresponding to the keywords contained in the voice recognition result. Model.
  • the above-mentioned three-dimensional model is a three-dimensional model inside a house.
  • the house includes a bedroom, a living room, a kitchen, and a bathroom, with a total of four sub-space scenes, that is, the above-mentioned three-dimensional model includes a three-dimensional sub-model of the bedroom, a three-dimensional sub-model of the living room, a three-dimensional sub-model of the kitchen, and a three-dimensional sub-model of the bathroom.
  • the keyword set includes bedroom, living room, kitchen, bathroom.
  • the keyword corresponding to the subspace scene bedroom can be "bedroom”; the keyword corresponding to the subspace scene kitchen can be “kitchen”; the keyword corresponding to the subspace scene living room can be “living room” ; The keyword corresponding to the bathroom in the subspace scene can be "toilet”. Further, as an example, if the voice recognition result includes the keyword "bedroom", then the second user terminal may present a three-dimensional sub-model of the bedroom on the aforementioned user interface.
  • the viewpoint switching of the three-dimensional model can be realized by voice, thereby presenting the subspace scene corresponding to the keywords contained in the voice recognition result.
  • the three-dimensional sub-model As a result, the convenience of browsing the three-dimensional model is improved, and the matching between the presented three-dimensional model and the voice acquired by the second user terminal is improved.
  • the present disclosure provides an embodiment of an interaction device based on a three-dimensional model.
  • the device embodiment may also include the same or corresponding features as the method embodiment shown in FIGS. 7, 9, and 10, and produce the same as those in the method embodiment shown in FIGS.
  • the method embodiments shown have the same or corresponding effects.
  • the interaction apparatus 1300 based on the three-dimensional model of this embodiment is set in a first user terminal, and the first user terminal presents a user interface.
  • the device 1300 includes: a first sending unit 1310 configured to send an interaction request for the target interaction operation to a server that provides page data for the user interface in response to detecting a user's target interaction operation on the user interface, wherein the user interface For presenting the three-dimensional model, the three-dimensional model is pre-associated with the user account logged in the second user terminal; the first receiving unit 1320 is configured to receive the streaming video obtained by the server from the second user terminal; the first presenting unit 1330 is It is configured to present streaming video and 3D models on the user interface.
  • the first sending unit 1310 of the interactive device 1300 based on the three-dimensional model may send the information about the target interaction operation to the server that provides page data for the user interface.
  • the user interface is used for presenting a three-dimensional model, and the three-dimensional model establishes an association relationship with the user account logged in by the second user terminal in advance.
  • the first receiving unit 1320 may receive the streaming video obtained by the server from the second user terminal.
  • the first presentation unit 1330 may present the streaming video and the three-dimensional model on the user interface.
  • the first receiving unit is further configured to: in response to the server receiving the interaction confirmation information sent by the second user terminal in response to the interaction request, receive the stream obtained by the server from the second user terminal. Media video.
  • the device 1300 further includes: a first adjustment unit (not shown in the figure) configured to respond to the current network speed value of the first user terminal being less than or equal to a preset Network speed threshold, based on each frame of voice in the streaming video, adjust the target user's image to generate a new video, where the new video characterizes the user indicated by the target user's image to perform the actions of each frame of voice instruction; the second presentation unit ( Figure Not shown in ), is configured to use new video instead of streaming video for presentation.
  • a first adjustment unit (not shown in the figure) configured to respond to the current network speed value of the first user terminal being less than or equal to a preset Network speed threshold, based on each frame of voice in the streaming video, adjust the target user's image to generate a new video, where the new video characterizes the user indicated by the target user's image to perform the actions of each frame of voice instruction
  • the second presentation unit ( Figure Not shown in ), is configured to use new video instead of streaming video for presentation.
  • the device 1300 further includes: a first generating unit (not shown in the figure) configured to generate a target user image based on an image in the streaming video; or A determining unit (not shown in the figure) is configured to determine the user image associated with the user account as the target user image.
  • the device 1300 further includes: a second sending unit (not shown in the figure), configured to send camera shutdown confirmation information to the server in response to a new video presented on the user interface , Wherein the camera closing confirmation information is used to determine whether the second user terminal closes the camera.
  • a second sending unit (not shown in the figure), configured to send camera shutdown confirmation information to the server in response to a new video presented on the user interface , Wherein the camera closing confirmation information is used to determine whether the second user terminal closes the camera.
  • the first receiving unit is further configured to: send the current network speed value of the first user terminal to the server; receive the streaming video that the server obtains and sends from the second user terminal , The streaming video has a resolution that matches the current network speed value.
  • the device 1300 further includes: a second receiving unit (not shown in the figure) configured to receive model adjustment information sent by the server, wherein the model adjustment information indicates the use of the first 2.
  • the user of the user terminal adjusts the three-dimensional model presented on the second user terminal.
  • the adjustment operation includes at least one of the following: zoom, rotate, move, and switch viewpoints; the second adjustment unit (not shown in the figure) is configured According to the adjustment operation indicated by the model adjustment information, the same adjustment operation is performed on the three-dimensional model presented on the user interface.
  • the three-dimensional model includes three-dimensional sub-models of multiple sub-space scenes, and the sub-space scenes in the multiple sub-space scenes correspond to keywords in a predetermined keyword set; and,
  • the device 1300 also includes: a first recognition unit (not shown in the figure), configured to perform voice recognition on the voice in the streaming video to obtain a voice recognition result; and a third presentation unit (not shown in the figure), configured In response to determining that the voice recognition result contains the keywords in the keyword set, a three-dimensional sub-model of the subspace scene corresponding to the keywords contained in the voice recognition result is presented on the user interface.
  • the device 1300 further includes: a first acquiring unit (not shown in the figure), configured to acquire user feedback information for streaming media videos; and a third sending unit (not shown in the figure). Not shown in ), is configured to send feedback information to a server, where the server is used to establish an association relationship between the feedback information and the user account.
  • the interaction device based on the three-dimensional model provided by the above-mentioned embodiments of the present disclosure is set in a first user terminal, and the first user terminal presents a user interface.
  • the first sending unit 1310 may send an interaction request for the target interaction operation to a server that provides page data for the user interface, where the user interface is used for The three-dimensional model is presented.
  • the three-dimensional model is pre-associated with the user account logged in the second user terminal.
  • the first receiving unit 1320 receives the streaming video obtained by the server from the second user terminal.
  • the first presenting unit 1330 displays on the user interface Streaming videos and 3D models are presented on the website.
  • streaming media videos and 3D models can be presented on the same page of the terminal device, which helps to use streaming media videos to present information related to the 3D model to users, which improves the diversity of interaction methods.
  • Users browse the 3D model more calmly, increase the user's browsing time, and help meet users' more diversified interactive needs.
  • the present disclosure provides an embodiment of a second interaction device based on a three-dimensional model.
  • the device embodiment may also include the same or corresponding features as the method embodiment shown in Figs. 11 and 12, and the method implementation shown in Figs. The same or corresponding effect.
  • the interaction device 1400 based on the three-dimensional model of this embodiment is set in the second user terminal.
  • the device 1400 includes: a second determining unit 1410 configured to obtain a streaming video in response to receiving an interaction request sent by a server, wherein the interaction request indicates that the first user terminal detects that the user interface presented to the first user terminal
  • the user interface is used to present a three-dimensional model, and the three-dimensional model has a pre-established association relationship with the user account logged in the second user terminal
  • the fourth sending unit 1420 is configured to send streaming video to the server, where the server It is used to send the streaming video to the first user terminal, so that the first user terminal presents the streaming video and the three-dimensional model on the user interface.
  • the second determining unit 1410 may obtain the streaming video.
  • the interaction request indicates that the first user terminal detects a user's target interaction operation on the user interface presented by the first user terminal, and the user interface is used to present a three-dimensional model.
  • the fourth sending unit 1420 may be configured to send the streaming video to the server, where the server is used to send the streaming video to the first user terminal, so that the first user terminal displays the streaming video on the user interface. And three-dimensional models.
  • the second determining unit 1410 is further configured to: in response to receiving the interaction request sent by the server, determine whether a confirmation operation of the user for the interaction request is detected; in response to detecting the confirmation Operation to obtain streaming video.
  • the device 1400 further includes: a third receiving unit (not shown in the figure), configured to respond to the current network speed value of the first user terminal being less than or equal to a preset The network speed threshold, receiving camera closing confirmation information from the server, and presenting camera closing confirmation information, where the camera closing confirmation information is used to determine whether the second user terminal closes the camera.
  • a third receiving unit (not shown in the figure), configured to respond to the current network speed value of the first user terminal being less than or equal to a preset The network speed threshold, receiving camera closing confirmation information from the server, and presenting camera closing confirmation information, where the camera closing confirmation information is used to determine whether the second user terminal closes the camera.
  • the device 1400 further includes: a fifth sending unit (not shown in the figure), configured to respond to receiving from the server the three-dimensional information presented on the second user terminal by the user.
  • the adjustment operation of the model is to perform the same adjustment operation on the three-dimensional model presented on the user interface according to the adjustment operation indicated by the model adjustment information, where the adjustment operation includes at least one of the following: zoom, rotate, move, and switch viewpoints.
  • the device 1400 further includes: a fifth sending unit (not shown in the figure), configured to respond to detecting that the user has received the three-dimensional model presented on the second user terminal
  • the adjustment operation is to send model adjustment information indicating the adjustment operation to the server, so that the server controls the first user terminal to perform the same adjustment operation on the three-dimensional model presented on the user interface according to the adjustment operation indicated by the model adjustment information.
  • the adjustment operation includes the following At least one item: zoom, rotate, move, viewpoint switch.
  • the three-dimensional model includes three-dimensional sub-models of multiple sub-space scenes, and the sub-space scenes in the multiple sub-space scenes correspond to keywords in a predetermined keyword set; and,
  • the device 1400 further includes: a second recognition unit (not shown in the figure), configured to perform voice recognition on the voice acquired by the first user terminal to obtain a voice recognition result; and a fourth presentation unit (not shown in the figure), It is configured to, in response to determining that the voice recognition result contains the keywords in the keyword set, present on the user interface a three-dimensional sub-model of the subspace scene corresponding to the keywords contained in the voice recognition result.
  • the apparatus 1400 further includes: an execution unit (not shown in the figure), configured to respond to receiving a stream media message sent by the server and sent by the user using the first user terminal. Video feedback information, perform operations that match the feedback information.
  • the interaction device based on the three-dimensional model provided by the above-mentioned embodiment of the present disclosure is set in the second user terminal, and the user account logged in by the second user terminal establishes an association relationship with the three-dimensional model in advance.
  • the second determining unit 1410 may obtain the streaming video, where the interaction request indicates that the first user terminal detects the user's target interaction operation on the user interface presented by the first user terminal, and the user interface is used to present the three-dimensional model.
  • the fourth sending unit 1420 may send the streaming video to the server, where the server is used to send the streaming video to the first user terminal, so that the first user terminal can present the streaming video and the three-dimensional model on the user interface.
  • streaming videos and 3D models can be presented on the same page of the terminal device, which helps to use streaming videos to present information related to the 3D model to users, which improves the diversity of interaction methods. Users browse the 3D model more calmly, increase the user's browsing time, and help meet users' more diversified interactive needs.
  • FIG. 15 is a schematic diagram of interaction of an embodiment 1500 of the interactive system based on a three-dimensional model of the present disclosure.
  • the interactive system based on the three-dimensional model includes a first user terminal, a second user terminal, and a server.
  • the first user terminal presents a user interface
  • the server is respectively communicatively connected with the first user terminal and the second user terminal.
  • the first user terminal, the second user terminal, and the server in the interactive system based on the three-dimensional model can perform the following steps:
  • Step 1501 The first user terminal detects the user's target interaction operation on the user interface.
  • the first user terminal detects the user's target interaction operation on the user interface.
  • the user interface is used for presenting a three-dimensional model, and the three-dimensional model establishes an association relationship with the user account logged in by the second user terminal in advance.
  • Step 1502 The first user terminal sends an interaction request for the target interaction operation to the server.
  • the first user terminal may send an interaction request for a target interaction operation to the server.
  • Step 1503 The second user terminal obtains the streaming video.
  • the second user terminal can obtain the streaming video.
  • Step 1504 The second user terminal sends the streaming video to the server.
  • the second user terminal may send the streaming video to the server.
  • Step 1505 The server sends the streaming video to the first user terminal.
  • the server may send the streaming video to the first user terminal.
  • Step 1506 The first user terminal presents the streaming video and the three-dimensional model on the user interface.
  • the first user terminal may present the streaming video and the three-dimensional model on the user interface.
  • step 1501 to step 1506 can also refer to the implementation of the first three-dimensional model-based interaction method described above.
  • the technical features in each embodiment of the second three-dimensional model-based interaction method, and the third three-dimensional model-based interaction method will be explained.
  • this embodiment may also include the same or corresponding features as the above-mentioned embodiment of the interaction method based on the three-dimensional model, and produce the same or corresponding effects, which will not be repeated here.
  • the interactive system based on the three-dimensional model provided by the above-mentioned embodiments of the present disclosure includes a first user terminal, a second user terminal, and a server.
  • the first user terminal presents a user interface
  • the server is in communication connection with the first user terminal and the second user terminal. .
  • the first user terminal is configured to: in response to detecting the user's target interaction operation on the user interface, send an interaction request for the target interaction operation to the server, wherein the user interface is used to present the three-dimensional model, the three-dimensional model and the second user
  • the user account logged in by the terminal establishes an association relationship in advance
  • the second user terminal is configured to: obtain the streaming video; send the streaming video to the server;
  • the server is also configured to: send the streaming video to the first user terminal;
  • the first user terminal It is configured to: present streaming video and three-dimensional model on the user interface.
  • streaming media videos and 3D models can be presented on the same page of the terminal device, which helps to use streaming media videos to present information related to the 3D model to users, which improves the diversity of interaction methods.
  • Users browse the 3D model more calmly, increase the user's browsing time, and help meet users' more diversified interactive needs.
  • FIG. 16 shows a block diagram of an electronic device 1600 according to an embodiment of the present disclosure.
  • the electronic device 1600 includes one or more processors 1611 and a memory 1612.
  • the processor 1611 may be a central processing unit (CPU) or another form of processing unit having the ability to implement three-dimensional scene interaction and/or instruction execution capabilities, and may control other components in the electronic device 1600 to perform desired functions .
  • CPU central processing unit
  • the memory 1612 may include one or more computer program products, and the computer program products may include various forms of computer-readable storage media, such as volatile memory and/or non-volatile memory.
  • the volatile memory for example, may include: random access memory (RAM) and/or cache memory (cache).
  • the non-volatile memory for example, may include: read-only memory (ROM), hard disk, flash memory, and the like.
  • One or more computer program instructions may be stored on the computer-readable storage medium, and the processor 1611 may run the program instructions to implement the various methods described above and/or other desired functions.
  • Various contents such as input signals, signal components, noise components, etc. can also be stored in the computer-readable storage medium.
  • the electronic device 1600 may further include: an input device 1613, an output device 1614, etc., and these components are interconnected by a bus system and/or other forms of connection mechanisms (not shown).
  • the input device 1613 may also include, for example, a keyboard, a mouse, and so on.
  • the output device 1614 can output various information to the outside.
  • the output device 1614 may include, for example, a display, a speaker, a printer, a communication network and a remote output device connected to it, and so on.
  • the electronic device 1600 may also include any other appropriate components.
  • the embodiments of the present disclosure may also be computer program products, which include computer program instructions that, when run by a processor, cause the processor to perform operations according to various embodiments of the present disclosure. Steps in various methods.
  • the computer program product may use any combination of one or more programming languages to write program codes for performing the operations of the embodiments of the present disclosure.
  • the programming languages include object-oriented programming languages, such as Java, C++, etc. , Also includes conventional procedural programming languages, such as "C" language or similar programming languages.
  • the program code can be executed entirely on the user's computing device, partly on the user's device, executed as an independent software package, partly on the user's computing device and partly executed on the remote computing device, or entirely on the remote computing device or server Executed on.
  • embodiments of the present disclosure may also be a computer-readable storage medium on which computer program instructions are stored.
  • the processor executes each of the various embodiments of the present disclosure. Steps in a method.
  • the computer-readable storage medium may adopt any combination of one or more readable media.
  • the readable medium may be a readable signal medium or a readable storage medium.
  • the readable storage medium may include, but is not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the above, for example. More specific examples (non-exhaustive list) of readable storage media may include: electrical connections with one or more wires, portable disks, hard disks, random access memory (RAM), read-only memory (ROM), Erasable programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
  • the method and apparatus of the present disclosure may be implemented in many ways.
  • the method and apparatus of the present disclosure can be implemented by software, hardware, firmware or any combination of software, hardware, and firmware.
  • the above-mentioned order of the steps for the method is for illustration only, and the steps of the method of the present disclosure are not limited to the order specifically described above, unless specifically stated otherwise.
  • the present disclosure can also be implemented as programs recorded in a recording medium, and these programs include machine-readable instructions for implementing methods according to embodiments of the present disclosure.
  • the present disclosure also covers a recording medium storing a program for executing a method according to an embodiment of the present disclosure.
  • each component or each step can be decomposed and/or recombined. These decomposition and/or recombination should be regarded as equivalent solutions of the present disclosure.

Abstract

A method for realizing interaction in a three-dimensional space scene comprises: responding to a user operation detected setting footprint information in a three-dimensional space scene, and determining a first pixel in a current view to which the user's current viewing angle in the three-dimensional space scene corresponds; determining a three-dimensional model to which the first pixel corresponds; determining the position of the user's footprint information in the three-dimensional model, wherein the footprint information is displayed when the three-dimensional space scene is being viewed; and setting the user's footprint information at the position.

Description

用于实现三维空间场景互动的方法、装置和设备Method, device and equipment for realizing three-dimensional space scene interaction 技术领域Technical field
本公开涉及虚拟现实全景技术和流媒体技术,尤其是涉及一种用于实现三维空间场景互动的方法、用于实现三维空间场景互动的装置、存储介质以及电子设备。The present disclosure relates to virtual reality panoramic technology and streaming media technology, and in particular to a method for realizing three-dimensional space scene interaction, a device for realizing three-dimensional space scene interaction, a storage medium, and electronic equipment.
背景技术Background technique
VR(Virtual Reality虚拟现实)全景技术是一种新兴的富媒体技术。由于VR全景技术可以720度无死角的为用户呈现三维空间场景,给用户带来浸入式视觉体验,因此,VR全景技术被广泛应用在网上商城、旅游服务以及房产服务等多种领域中。如何使VR全景技术给用户带来更丰富的使用体验,是一个值得关注的技术问题。VR (Virtual Reality) panoramic technology is an emerging rich media technology. Because VR panoramic technology can present users with three-dimensional space scenes without blind angles at 720 degrees, and bring users an immersive visual experience, VR panoramic technology is widely used in various fields such as online shopping malls, travel services, and real estate services. How to enable VR panoramic technology to bring users a richer experience is a technical issue worthy of attention.
相对于二维画面,三维模型可以给人以更为强烈的视觉观感。有了物体的三维数据,便可以向用户呈现该物体的任意视图,并且,视图间可以保持正确的投影关系。Compared with two-dimensional images, three-dimensional models can give people a stronger visual perception. With the three-dimensional data of the object, any view of the object can be presented to the user, and the correct projection relationship can be maintained between the views.
现有技术中,在用户终端呈现三维模型的同时,可以支持用户终端间的实时语音同屏交互,也即,在用户终端呈现三维模型的过程中,用户终端的对端用户的语音可以实时传输至该用户终端,并且,用户终端获取的语音也可以实时传输至对端。In the prior art, while the user terminal is presenting a three-dimensional model, it can support real-time voice on-screen interaction between user terminals, that is, in the process of the user terminal presenting the three-dimensional model, the voice of the opposite user of the user terminal can be transmitted in real time To the user terminal, and the voice acquired by the user terminal can also be transmitted to the opposite terminal in real time.
然而,现有技术中的上述交互方式较为单一,语音交互通常存在局限性。However, the foregoing interaction methods in the prior art are relatively single, and voice interaction usually has limitations.
发明内容Summary of the invention
根据本公开实施例的一个方面,提供了一种用于实现三维空间场景互动的方法,包括:响应于检测到在三维空间场景中设置足迹信息的用户操作,确定用户在三维空间场景中的当前视角所对应的当前视图中的第一像素点;确定第一像素点对应的三维模型;确定用户的足迹信息在所述三维模型中的位置,其中足迹信息用于在三维空间场景被浏览时显示;以及在位置处设置用户的足迹信息。According to one aspect of the embodiments of the present disclosure, there is provided a method for realizing interaction in a three-dimensional space scene, including: in response to detecting a user operation of setting footprint information in the three-dimensional space scene, determining the user's current position in the three-dimensional space scene The first pixel in the current view corresponding to the perspective; determine the three-dimensional model corresponding to the first pixel; determine the position of the user's footprint information in the three-dimensional model, where the footprint information is used to display when the three-dimensional space scene is viewed ; And set the user's footprint information at the location.
根据本公开实施例的另一个方面,提供了一种基于三维模型的交互方法,包括:在呈现有用户界面的第一用户终端处:响应于检测到用户针对用户界面的目标交互操作,向为用户界面提供页面数据的服务器发送针对目标交互操作的交互请求,其中,用户界面用于呈现三维模型,三维模型与第二用户终端登录的用户账号预先建立关联关系;接收所述服务器从第二用户终端获取的流媒体视频;以及在用户界面上呈现流媒体视频和三维模型。According to another aspect of the embodiments of the present disclosure, there is provided an interaction method based on a three-dimensional model, including: at a first user terminal presenting a user interface: in response to detecting a user's target interaction operation on the user interface, The server that provides page data in the user interface sends an interaction request for the target interaction operation, where the user interface is used to present a three-dimensional model, and the three-dimensional model establishes an association relationship with the user account logged in the second user terminal in advance; Streaming video obtained by the terminal; and presenting the streaming video and three-dimensional model on the user interface.
根据本公开实施例的另一方面,提供了一种基于三维模型的交互方法,包括:在第二用户终端处:响应于接收到服务器发送的交互请求,获取流媒体视频,其中,交互请求指示第一用户终端检测到用户针对第一用户终端呈现的用户界面的目标交互操作,用户界面用于呈现三维模型,三维模型与第二用户终端登录的用户账号预先建立关联关系;以及向服务器发送流媒体视频,其中,服务器用于将流媒体视频发送至所述第一用户终端,以使第一用户终端在用户界面上呈现流媒体视频和三维模型。According to another aspect of the embodiments of the present disclosure, there is provided an interaction method based on a three-dimensional model, including: at a second user terminal: in response to receiving an interaction request sent by a server, acquiring a streaming video, wherein the interaction request indicates The first user terminal detects the user's target interaction operation on the user interface presented by the first user terminal, the user interface is used to present a three-dimensional model, and the three-dimensional model establishes a pre-association relationship with the user account logged in by the second user terminal; and sends a stream to the server Media video, where the server is used to send the streaming video to the first user terminal, so that the first user terminal presents the streaming video and the three-dimensional model on the user interface.
根据本公开实施例的另一个方面,提供了一种用于实现三维空间场景互动的装置,包括:用于执行上述方法中任一项所述的方法的装置。According to another aspect of the embodiments of the present disclosure, there is provided a device for realizing interaction of a three-dimensional space scene, including: a device for executing the method described in any one of the above methods.
根据本公开实施例的另一方面,提供了一种基于三维模型的交互装置,设置于第一用户终端,装置包括:用于执行上述方法中任一项所述的方法的装置。According to another aspect of the embodiments of the present disclosure, there is provided an interaction device based on a three-dimensional model, which is provided in a first user terminal, and the device includes: a device for executing the method described in any one of the above methods.
根据本公开实施例的另一方面,提供了一种基于三维模型的交互装置,设置于第二用户终端,装置包括:用于执行上述方法中中任一项所述的方法的装置。According to another aspect of the embodiments of the present disclosure, there is provided an interaction device based on a three-dimensional model, which is provided in a second user terminal, and the device includes: a device for executing the method described in any one of the above methods.
根据本公开实施例的另一方面,提供了一种基于三维模型的交互系统,包括:第一用户终端,用于呈现用户界面;第二用户终端;和服务器,服务器与第一用户终端和第二用户终端通信连接,第一用户终端被配置成:响应于检测到用户针对用户界面的目标交互操作,向服务器发送针对目标交互操作的交互请求,用户界面用于呈现三维模型,三维模型与第二用户终端登录的用户账号预先建立关联关系;第二用户终端被配置成:获取流媒体视频;并且向服务器发送所述流媒体视频;服务器被配置成:向第一用户终端发送所述流媒体视频;并且第一用户终端被配置成:在用户界面上呈现流媒体视频和三维模型。According to another aspect of the embodiments of the present disclosure, there is provided an interactive system based on a three-dimensional model, including: a first user terminal for presenting a user interface; a second user terminal; and a server. The second user terminal is in communication connection. The first user terminal is configured to: in response to detecting the user's target interaction operation on the user interface, send an interaction request for the target interaction operation to the server, and the user interface is used to present the three-dimensional model, the three-dimensional model and the second 2. The user account logged in by the user terminal establishes an association relationship in advance; the second user terminal is configured to: obtain the streaming video; and send the streaming video to the server; the server is configured to: send the streaming media to the first user terminal Video; and the first user terminal is configured to: present a streaming video and a three-dimensional model on the user interface.
根据本公开实施例的另一个方面,提供了一种非暂态性计算机可读存储介质,存储介质存储有计算机程序,计算机程序当被计算机执行时使计算机实现上述方法中任一项所述的方法。According to another aspect of the embodiments of the present disclosure, a non-transitory computer-readable storage medium is provided, and the storage medium stores a computer program. method.
根据本公开实施例的另一个方面,提供了一种电子设备,包括:处理器;以及用于存储处理器可执行指令的存储器,处理器可执行指令在被处理器执行时实现上述方法中任一项所述的方法。According to another aspect of the embodiments of the present disclosure, there is provided an electronic device including: a processor; and a memory for storing processor-executable instructions. The processor-executable instructions implement any of the above methods when executed by the processor. The method described in one item.
根据本公开的实施例的另一个方面,提供了一种计算机程序产品,包括计算机程序,计算机程序当被计算机执行时使计算机实现上述方法中任一项所述的方法。According to another aspect of the embodiments of the present disclosure, there is provided a computer program product, including a computer program, which when executed by a computer causes the computer to implement the method described in any one of the above methods.
下面通过附图和实施例,对本公开的技术方案做进一步的详细描述。The technical solutions of the present disclosure will be further described in detail below through the accompanying drawings and embodiments.
附图说明Description of the drawings
构成说明书的一部分的附图描述了本公开的实施例,并且连同描述一起用于解释本公开的 原理。The drawings constituting a part of the specification describe the embodiments of the present disclosure, and together with the description, serve to explain the principle of the present disclosure.
参照附图,根据下面的详细描述,可以更加清楚地理解本公开,其中:With reference to the accompanying drawings, the present disclosure can be understood more clearly according to the following detailed description, in which:
图1为本公开的适用场景的一个实施例的示意图;FIG. 1 is a schematic diagram of an embodiment of an applicable scenario of the present disclosure;
图2为本公开的用于实现三维空间场景互动的方法一个实施例的流程图;FIG. 2 is a flowchart of an embodiment of a method for realizing interaction of a three-dimensional space scene of the present disclosure;
图3为本公开的确定第一像素点对应的三维模型一实施例的流程图;3 is a flowchart of an embodiment of determining a three-dimensional model corresponding to a first pixel in the present disclosure;
图4为本公开的确定第一像素点对应的三维模型另一实施例的流程图;4 is a flowchart of another embodiment of determining a three-dimensional model corresponding to a first pixel point of the present disclosure;
图5为本公开的为浏览用户呈现足迹信息一实施例的流程图;FIG. 5 is a flowchart of an embodiment of presenting footprint information for browsing users in the present disclosure;
图6为本公开的用于实现三维空间场景互动的装置一个实施例的结构示意图;FIG. 6 is a schematic structural diagram of an embodiment of an apparatus for realizing interaction in a three-dimensional space scene of the present disclosure;
图7为本公开的第一个基于三维模型的交互方法的一个实施例的流程图。FIG. 7 is a flowchart of an embodiment of the first interaction method based on a three-dimensional model of the present disclosure.
图8A-图8C是针对图7的实施例的应用场景示意图。8A-8C are schematic diagrams of application scenarios for the embodiment of FIG. 7.
图9为本公开的第一个基于三维模型的交互方法的另一个实施例的流程图。Fig. 9 is a flowchart of another embodiment of the first three-dimensional model-based interaction method of the present disclosure.
图10为本公开的第一个基于三维模型的交互方法的又一个实施例的流程图。FIG. 10 is a flowchart of another embodiment of the first interaction method based on a three-dimensional model of the present disclosure.
图11为本公开的第二个基于三维模型的交互方法的一个实施例的流程图。FIG. 11 is a flowchart of an embodiment of the second three-dimensional model-based interaction method of the present disclosure.
图12为本公开的第二个基于三维模型的交互方法的另一个实施例的流程图。FIG. 12 is a flowchart of another embodiment of the second three-dimensional model-based interaction method of the present disclosure.
图13为本公开的第一个基于三维模型的交互装置的一个实施例的流程图。FIG. 13 is a flowchart of an embodiment of the first interactive device based on a three-dimensional model of the present disclosure.
图14为本公开的第二个基于三维模型的交互装置的一个实施例的流程图。FIG. 14 is a flowchart of an embodiment of the second interactive device based on a three-dimensional model of the present disclosure.
图15为本公开的基于三维模型的交互系统的一个实施例的交互示意图。FIG. 15 is a schematic diagram of interaction of an embodiment of the interactive system based on a three-dimensional model of the present disclosure.
图16为本公开一示例性实施例提供的电子设备的结构图。Fig. 16 is a structural diagram of an electronic device provided by an exemplary embodiment of the present disclosure.
具体实施方式Detailed ways
下面将参考附图详细地描述根据本公开的示例实施例。显然,所描述的实施例仅仅是本公开的一部分实施例,而不是本公开的全部实施例,应理解,本公开不受这里描述的示例实施例的限制。Hereinafter, exemplary embodiments according to the present disclosure will be described in detail with reference to the accompanying drawings. Obviously, the described embodiments are only a part of the embodiments of the present disclosure, rather than all the embodiments of the present disclosure, and it should be understood that the present disclosure is not limited by the exemplary embodiments described herein.
应注意到:除非另外具体说明,否则在这些实施例中阐述的部件和步骤的相对布置、数字表达式和数值不限制本公开的范围。It should be noted that unless specifically stated otherwise, the relative arrangement of components and steps, numerical expressions and numerical values set forth in these embodiments do not limit the scope of the present disclosure.
本领域技术人员可以理解,本公开实施例中的“第一”、“第二”等术语仅用于区别不同步骤、设备或模块等,既不代表任何特定技术含义,也不表示它们之间的必然逻辑顺序。Those skilled in the art can understand that terms such as “first” and “second” in the embodiments of the present disclosure are only used to distinguish different steps, devices, or modules, etc., and do not represent any specific technical meaning, nor do they mean any difference between them. The necessary logical order.
还应理解,在本公开实施例中,“多个”可以指两个或者两个以上,“至少一个”可以指一个、两个或两个以上。It should also be understood that in the embodiments of the present disclosure, "plurality" may refer to two or more than two, and "at least one" may refer to one, two, or more than two.
还应理解,对于本公开实施例中提及的任一部件、数据或结构,在没有明确限定或者在前 后文给出相反启示的情况下,一般可以理解为一个或多个。It should also be understood that any component, data, or structure mentioned in the embodiments of the present disclosure can generally be understood as one or more unless it is clearly defined or given opposite enlightenment.
另外,本公开中术语“和/或”,仅是一种描述关联对象的关联关系,表示可以存在三种关系,如A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。另外,本公开中字符“/”,一般表示前后关联对象是一种“或”的关系。In addition, the term "and/or" in the present disclosure is only an association relationship that describes associated objects, indicating that there can be three relationships, such as A and/or B, which can mean: A alone exists, and A and B exist at the same time. There are three cases of B alone. In addition, the character "/" in the present disclosure generally indicates that the associated objects before and after are in an "or" relationship.
还应理解,本公开对各个实施例的描述着重强调各个实施例之间的不同之处,其相同或相似之处可以相互参考,为了简洁,不再一一赘述。It should also be understood that the description of the various embodiments of the present disclosure emphasizes the differences between the various embodiments, and the same or similarities can be referred to each other, and for the sake of brevity, they will not be repeated one by one.
同时,应当明白,为了便于描述,附图中所示出的各个部分的尺寸并不是按照实际的比例关系绘制的。At the same time, it should be understood that, for ease of description, the sizes of the various parts shown in the drawings are not drawn according to actual proportional relationships.
以下对至少一个示例性实施例的描述实际上仅仅是说明性的,决不作为对本公开及其应用或使用的任何限制。The following description of at least one exemplary embodiment is actually only illustrative, and in no way serves as any limitation to the present disclosure and its application or use.
对于相关领域普通技术人员已知的技术、方法和设备可能不作详细讨论,但在适当情况下,所述技术、方法和设备应当被视为说明书的一部分。The technologies, methods, and equipment known to those of ordinary skill in the relevant fields may not be discussed in detail, but where appropriate, the technologies, methods, and equipment should be regarded as part of the specification.
应注意到:相似的标号和字母在下面的附图中表示类似项,因此,一旦某一项在一个附图中被定义,则在随后的附图中不需要对其进行进一步讨论。It should be noted that similar reference numerals and letters indicate similar items in the following drawings, therefore, once an item is defined in one drawing, it does not need to be further discussed in the subsequent drawings.
本公开的实施例可以应用于终端设备、计算机系统、服务器等电子设备,其可与众多其它通用或者专用计算系统环境或配置一起操作。适于与终端设备、计算机系统或者服务器等电子设备一起使用的众所周知的终端设备、计算系统、环境和/或配置的例子包括但不限于:个人计算机系统、服务器计算机系统、瘦客户机、厚客户机、手持或膝上设备、基于微处理器的系统、机顶盒、可编程消费电子产品、网络个人电脑、小型计算机系统、大型计算机系统和包括上述任何系统的分布式云计算技术环境等等。The embodiments of the present disclosure can be applied to electronic devices such as terminal devices, computer systems, servers, etc., which can operate with many other general-purpose or special-purpose computing system environments or configurations. Examples of well-known terminal devices, computing systems, environments and/or configurations suitable for use with electronic devices such as terminal devices, computer systems, or servers include, but are not limited to: personal computer systems, server computer systems, thin clients, and thick clients Computers, handheld or laptop devices, microprocessor-based systems, set-top boxes, programmable consumer electronics, network personal computers, small computer systems, large computer systems, and distributed cloud computing technology environments including any of the above systems, etc.
终端设备、计算机系统、服务器等电子设备可以在由计算机系统执行的计算机系统可执行指令(诸如程序模块)的一般语境下描述。通常,程序模块可以包括例程、程序、目标程序、组件、逻辑、数据结构等等,它们执行特定的任务或者实现特定的抽象数据类型。计算机系统/服务器可以在分布式云计算环境中实施。在分布式云计算环境中,任务可以是由通过通信网络链接的远程处理设备执行的。在分布式云计算环境中,程序模块可以位于包括存储设备的本地或远程计算系统存储介质上。Electronic devices such as terminal devices, computer systems, and servers can be described in the general context of computer system executable instructions (such as program modules) executed by the computer system. Generally, program modules may include routines, programs, object programs, components, logic, data structures, etc., which perform specific tasks or implement specific abstract data types. The computer system/server can be implemented in a distributed cloud computing environment. In a distributed cloud computing environment, tasks can be performed by remote processing devices linked through a communication network. In a distributed cloud computing environment, program modules may be located on a storage medium of a local or remote computing system including a storage device.
在实现本公开的过程中,发明人发现,在用户通过调整其当前视角来感受三维空间场景的过程中,往往会产生一些情绪以及想法等感受。如果用户能够将用于表征其感受的足迹信息设置到三维空间场景中,则不仅有利于提高用户自身的参与感,而且,用户留下的足迹信息还可以为观看该三维空间场景的其他用户带来更丰富的VR全景体验。In the process of realizing the present disclosure, the inventor found that in the process of a user experiencing a three-dimensional space scene by adjusting his current perspective, some feelings such as emotions and thoughts are often generated. If the user can set the footprint information used to characterize his feelings into the three-dimensional space scene, it will not only help improve the user’s sense of participation, but the footprint information left by the user can also provide other users watching the three-dimensional space scene. Come for a richer VR panoramic experience.
下面结合图1,对本公开提供的用于实现三维空间场景互动的技术的应用场景的一个例子,进行说明。In the following, an example of an application scenario of the technology for realizing three-dimensional space scene interaction provided by the present disclosure will be described with reference to FIG. 1.
在房产领域,可以利用VR全景技术为待出租房屋或者待出售房屋,设置三维空间场景。任一用户均可以通过网络访问,随时随地的观看相应房屋的三维空间场景。在用户观看相应房屋的三维空间场景的过程中,本公开允许该用户针对其所浏览的房屋留下自己的足迹信息,而且,本公开可以将该用户自己针对该房屋留下的足迹信息和其他用户针对该房屋留下的足迹信息,一并呈现给该用户。In the real estate field, VR panoramic technology can be used to set a three-dimensional space scene for a house to be rented or a house to be sold. Any user can access through the network and watch the three-dimensional space scene of the corresponding house anytime and anywhere. In the process of the user watching the three-dimensional space scene of the corresponding house, the present disclosure allows the user to leave his own footprint information for the house he is browsing, and the present disclosure can target the user’s own footprint information and other information about the house. The footprint information left by the user for the house is presented to the user together.
在一个具体的例子中,假设用户正在浏览房屋的三维空间场景,基于该用户的当前视角所看到的当前视图如图1所示。In a specific example, suppose that the user is browsing the three-dimensional space scene of the house, and the current view based on the user's current perspective is shown in FIG. 1.
其他用户针对该两室一厅的房屋的三维空间场景留下的足迹信息120包括:“喜欢这组沙发,棒棒哒”、“这个装饰隔断不错”、“这个沙发不错,高级上档次”、“组合搭配很用心,赞赞赞”、“茶几造型设计的很独特~文案最长二十个字啊”以及图1右上角所示的三维模型110。为浏览该房屋的三维空间场景的用户呈现其他用户针对该房屋的三维空间场景所留下的足迹信息120,有利于使该用户了解其他用户对该房屋的感受,从而有利于加深用户对该房屋的认知,进而有利于提高用户对该房屋的浏览体验。The footprint information 120 left by other users for the three-dimensional space scene of the two-bedroom and one-living house includes: "I like this group of sofas, it's great", "This decorative partition is good", "This sofa is good, high-end class", "Combination and collocation are very diligent, praise and praise", "The design of the tea table is very unique ~ the longest copywriting is 20 characters" and the three-dimensional model 110 shown in the upper right corner of Figure 1. The user who browses the three-dimensional space scene of the house is presented with the footprint information 120 left by other users with respect to the three-dimensional space scene of the house, which helps the user understand other users’ feelings about the house, thereby helping to deepen the user’s experience of the house. Cognition, which in turn helps to improve the user’s browsing experience of the house.
另外,用户在观看该房屋的三维空间场景的过程中,也可以发表自己对该房屋的感受,即在该三维空间场景中留下自己的足迹信息。例如,用户可以在图1所示的柱子的位置处设置“这个柱子让房屋看起来更具特色”等足迹信息。用户自己设置的足迹信息可以即时显示在图1所示的三维空间场景中,即用户可以在观看该房屋的三维空间场景的过程中看到自己留下的足迹信息,从而有利于提升用户的参与感。In addition, in the process of watching the three-dimensional space scene of the house, the user can also express his own feelings about the house, that is, leave his own footprint information in the three-dimensional space scene. For example, the user can set footstep information such as "this pillar makes the house look more distinctive" at the position of the pillar shown in FIG. The footprint information set by the user can be instantly displayed in the three-dimensional space scene shown in Figure 1, that is, the user can see the footprint information left by himself during the process of viewing the three-dimensional space scene of the house, which is conducive to enhancing user participation feel.
再有,所有用户针对该房屋设置的不属于图1所示的视图的其他足迹信息,可以以弹幕130的形式呈现给用户,从而有利于提升用户浏览该房屋的其他位置的三维空间场景的兴趣。Furthermore, all other footprint information set by the user for the house that does not belong to the view shown in FIG. 1 can be presented to the user in the form of a bullet screen 130, which is beneficial to improve the user's ability to browse three-dimensional space scenes in other locations of the house. interest.
本公开提供的用于实现三维空间场景互动的技术还可以应用在其他多种场景中,例如,用户在浏览图书馆的三维空间场景时,可以针对图书馆中的一本书或者座椅或者咖啡机等设置相应的足迹信息。用户针对书设置的足迹信息可以是用户针对书的观感或者用户当前阅读到的页数等。在此对本公开提供的用于实现三维空间场景互动的技术所能应用的场景,不再一一举例说明。The technology for realizing the interaction of three-dimensional space scenes provided by the present disclosure can also be applied to various other scenes. For example, when a user browses a three-dimensional space scene of a library, it can target a book or chair or coffee in the library. Set the corresponding footprint information. The footprint information set by the user for the book may be the user's impression of the book or the number of pages currently read by the user. Here, the scenarios where the technology for realizing the interaction of three-dimensional space scenes provided by the present disclosure can be applied will not be illustrated one by one.
图2为本公开的用于实现三维空间场景互动的方法一个实施例的流程图。图2所示的实施例的方法200包括步骤210至240。下面对各步骤分别进行说明。FIG. 2 is a flowchart of an embodiment of a method for realizing interaction of a three-dimensional space scene of the present disclosure. The method 200 of the embodiment shown in FIG. 2 includes steps 210 to 240. Each step is described separately below.
在步骤210,响应于检测到在三维空间场景中设置足迹信息的用户操作,确定用户在该三 维空间场景中的当前视角所对应的当前视图中的第一像素点。In step 210, in response to detecting the user operation of setting the footprint information in the three-dimensional space scene, determine the first pixel in the current view corresponding to the user's current perspective in the three-dimensional space scene.
根据本公开的示范性实施例,三维空间场景可以是指利用预先设置的全景图和三维模型,为用户呈现出的具有三维立体感的空间场景。例如,三维空间场景可以为针对图书馆设置的三维空间场景、针对房屋设置的三维空间场景、针对咖啡馆设置的三维空间场景、或者针对商场设置的三维空间场景等。According to an exemplary embodiment of the present disclosure, a three-dimensional space scene may refer to a space scene with a three-dimensional sense that is presented to the user by using a preset panoramic image and a three-dimensional model. For example, the three-dimensional space scene may be a three-dimensional space scene set for a library, a three-dimensional space scene set for a house, a three-dimensional space scene set for a cafe, or a three-dimensional space scene set for a shopping mall.
根据本公开的示范性实施例,可以在用户触发在三维空间场景中设置足迹信息的功能时,检测到用户需要在该三维空间场景中设置足迹信息。例如,在用户点击用于设置足迹信息的按钮或者菜单上的相应选项时,本公开的实施例可以检测到用户需要在三维空间场景中设置足迹信息。再例如,用户可以使用预先设置的快捷方式触发在三维空间场景中设置足迹信息的功能。本公开的实施例中,用户的足迹信息可以为能够表示出用户曾经到访过该三维空间场景的信息。该足迹信息可以认为是用户的来访痕迹信息。According to an exemplary embodiment of the present disclosure, when the user triggers the function of setting the footprint information in the three-dimensional space scene, it can be detected that the user needs to set the footprint information in the three-dimensional space scene. For example, when the user clicks a button for setting footstep information or a corresponding option on a menu, the embodiment of the present disclosure can detect that the user needs to set footstep information in a three-dimensional space scene. For another example, the user can use a preset shortcut to trigger the function of setting footprint information in the three-dimensional space scene. In the embodiment of the present disclosure, the user's footprint information may be information that can indicate that the user has visited the three-dimensional space scene. The footprint information can be considered as the visit trace information of the user.
根据本公开的示范性实施例,用户在三维空间场景中的当前视角可以是指:用户当前观看三维空间场景的位置和角度。用户在三维空间场景中的当前视角通常会随着用户的操作,而发生变化。例如,用户通过在触摸屏上执行拖拽等操作,可以控制其在三维空间场景中的当前视角。用户在三维空间场景中的当前视角决定了用户当前所能看到的全景图的内容/区域,即用户在三维空间场景中的当前视角决定了当前视图。According to an exemplary embodiment of the present disclosure, the current perspective of the user in the three-dimensional space scene may refer to the position and angle at which the user currently views the three-dimensional space scene. The user's current perspective in the three-dimensional space scene usually changes with the user's operation. For example, the user can control his current perspective in the three-dimensional scene by performing operations such as dragging on the touch screen. The user's current perspective in the three-dimensional space scene determines the content/area of the panorama that the user can currently see, that is, the user's current perspective in the three-dimensional space scene determines the current view.
根据本公开的示范性实施例,第一像素点为当前视图中的一个像素点。可以根据预先设置的缺省规则,获得第一像素点。例如,第一像素点可以为当前视图中的一个特定像素点,也可以为当前视图中的任意一个像素点。According to an exemplary embodiment of the present disclosure, the first pixel point is one pixel point in the current view. The first pixel point can be obtained according to a preset default rule. For example, the first pixel may be a specific pixel in the current view, or it may be any pixel in the current view.
在步骤220,确定该第一像素点对应的三维模型。In step 220, the three-dimensional model corresponding to the first pixel is determined.
根据本公开的示范性实施例,三维空间场景通常由多个三维模型形成。在一些实施例中,三维空间场景也可以由一个三维模型形成。用户看到的当前视图中的一个像素点可以是三维模型中的一个点的呈现。用户看到的当前视图中的一个像素点也可以不是三维模型中的任一点的呈现。也就是说,在通常情况下,三维空间场景中的任一三维模型中的任一点均可以呈现在全景图中,而全景图中的点所呈现的可能并不是三维空间场景中的所有三维模型中的点。当然,本公开也不排除三维空间场景中的三维模型中的部分点未被呈现在全景图中的可能性。According to an exemplary embodiment of the present disclosure, a three-dimensional space scene is generally formed by a plurality of three-dimensional models. In some embodiments, the three-dimensional space scene may also be formed by a three-dimensional model. A pixel point in the current view seen by the user may be a representation of a point in the three-dimensional model. A pixel in the current view that the user sees may not be a representation of any point in the three-dimensional model. That is to say, under normal circumstances, any point in any three-dimensional model in the three-dimensional space scene can be presented in the panorama, and the points in the panoramic image may not be all three-dimensional models in the three-dimensional scene. In the point. Of course, the present disclosure does not exclude the possibility that some points in the three-dimensional model in the three-dimensional space scene are not presented in the panoramic image.
在一些示范性实施例中,在第一像素点用于呈现三维模型中的点时,该点所在的三维模型即为第一像素点对应的三维模型。In some exemplary embodiments, when the first pixel is used to present a point in the three-dimensional model, the three-dimensional model where the point is located is the three-dimensional model corresponding to the first pixel.
在一些示范性实施例中,在第一像素点用于呈现非三维模型中的点时,可以利用当前视图中的其他像素点来更新第一像素点。在一些实施例中,也可以不对第一像素点进行更新处理, 此时,第一像素点对应的三维模型可以为:当前视图中与第一像素点相近的用于呈现三维模型中的点的其他像素点所对应的三维模型。也就是说,在第一像素点用于呈现非三维模型中的点,且不更新第一像素点的情况下,可以将当前视图中其他像素点对应的三维模型作为第一像素点对应的三维模型。In some exemplary embodiments, when the first pixel is used to present a point in a non-three-dimensional model, other pixels in the current view may be used to update the first pixel. In some embodiments, the first pixel may not be updated. At this time, the three-dimensional model corresponding to the first pixel may be: the current view that is similar to the first pixel and is used to present the points in the three-dimensional model. The three-dimensional model corresponding to other pixels. That is to say, when the first pixel is used to present the points in the non-three-dimensional model, and the first pixel is not updated, the three-dimensional model corresponding to other pixels in the current view can be used as the three-dimensional model corresponding to the first pixel. Model.
在步骤230,确定该用户的足迹信息在该三维模型中的位置,其中足迹信息用于在三维空间场景被浏览时显示。In step 230, the position of the user's footprint information in the three-dimensional model is determined, where the footprint information is used to display when the three-dimensional space scene is browsed.
由于全景图中的至少部分像素点与三维模型中的点存在映射关系,因此,可以获得上述第一像素点或者上述其他像素点在三维模型中的位置。该位置即为用户的足迹信息的位置。Since at least some of the pixels in the panoramic image have a mapping relationship with the points in the three-dimensional model, the positions of the first pixel or the other pixels in the three-dimensional model can be obtained. This location is the location of the user's footprint information.
根据本公开的示范性实施例,三维空间场景中的所有三维模型可以分别设置有各自的三维坐标系,也可以具有同一的三维坐标系。用户的足迹信息在三维模型中的位置可以通过(x,y,z)来表示。即,用户的足迹信息可以是具有深度的。According to an exemplary embodiment of the present disclosure, all three-dimensional models in a three-dimensional space scene may be respectively provided with their own three-dimensional coordinate systems, or may have the same three-dimensional coordinate system. The position of the user's footprint information in the three-dimensional model can be represented by (x, y, z). That is, the user's footprint information can be deep.
在步骤240,在位置处设置用户的足迹信息。In step 240, the user's footprint information is set at the location.
根据本公开的示范性实施例,在该位置处设置该用户的足迹信息可以包括:为该用户的足迹信息设置三维模型标识以及三维坐标,并存储三维模型标识、三维坐标以及用户的足迹信息的对应关系。According to an exemplary embodiment of the present disclosure, setting the user's footprint information at the location may include: setting a three-dimensional model identifier and three-dimensional coordinates for the user's footprint information, and storing the three-dimensional model identifier, three-dimensional coordinates, and user's footprint information. Correspondence.
根据本公开的示范性实施例,用户的足迹信息可以用于显示给该三维空间场景的浏览用户(如所有浏览用户或者部分浏览用户)。本公开的实施例中,三维空间场景的浏览用户可以包括设置该足迹信息的用户。According to an exemplary embodiment of the present disclosure, the user's footprint information may be used to display to browsing users (such as all browsing users or some browsing users) of the three-dimensional space scene. In the embodiment of the present disclosure, the browsing user of the three-dimensional space scene may include the user who sets the footprint information.
根据本公开的示范性实施例,通过利用需要设置足迹信息的用户的当前视图中的第一像素点,获得第一像素点对应的三维模型以及足迹信息在该三维模型中的位置,使用户设置的足迹信息可以与相应三维模型的相应位置产生关联。这样,在基于用户的当前视角,利用全景图来呈现三维模型,以形成三维空间场景时,可以使用户的足迹信息呈现在三维空间场景的适当位置处,从而有利于使用户对三维空间场景中的具体部位的感受,精准的呈现在三维空间场景的相应位置处。由此,实现了用户与三维空间场景的互动。这不仅有利于提高用户自身的参与感以及沉浸感,并提升用户在三维空间场景中的停留时长,而且,用户留下的足迹信息还可以为该三维空间场景的至少一浏览用户带来更丰富的VR全景体验。According to an exemplary embodiment of the present disclosure, by using the first pixel in the current view of the user who needs to set the footprint information, the three-dimensional model corresponding to the first pixel and the position of the footprint information in the three-dimensional model are obtained, so that the user can set The footprint information of can be associated with the corresponding position of the corresponding 3D model. In this way, when a panoramic view is used to present a three-dimensional model based on the user’s current perspective to form a three-dimensional space scene, the user’s footprint information can be presented at an appropriate position in the three-dimensional space scene, which is conducive to the user’s understanding of the three-dimensional space scene. The feelings of the specific parts are accurately presented in the corresponding position of the three-dimensional space scene. In this way, the interaction between the user and the three-dimensional space scene is realized. This is not only conducive to improving the user’s sense of participation and immersion, and increasing the length of time the user stays in the 3D space scene, but the footprint information left by the user can also bring more enrichment to at least one user who browses the 3D space scene. VR panoramic experience.
在一个可选示例中,足迹信息包括:文本、图片、音频、视频以及三维模型中的至少一个。文本可以认为是字符(如文字、字母、数字或者符号等)形式的留言。图片可以认为是图像(如照片或者表情符号等)形式的留言。音频可以认为是声音形式的留言(也可以称为留音等)。视频可以认为是影像形式的留言。三维模型可以认为是立体形式的留言。本公开的实施例中, 用户的足迹信息可以称为用户的留言。用户设置的一条足迹信息可以同时包括:文本、图片、音频、视频以及三维模型中的一种或者多种。通过使用户的足迹信息包括文本、图片、音频、视频以及三维模型中的至少一个,有利于使用户的足迹信息的表现形式更加丰富,从而有利于丰富用户与三维空间场景的互动方式。In an optional example, the footprint information includes: at least one of text, picture, audio, video, and a three-dimensional model. The text can be considered as a message in the form of characters (such as text, letters, numbers, or symbols, etc.). A picture can be considered as a message in the form of an image (such as a photo or emoticon, etc.). Audio can be thought of as a voice message (also called memo, etc.). Video can be thought of as a message in the form of an image. The three-dimensional model can be considered as a three-dimensional message. In the embodiments of the present disclosure, the user's footprint information may be referred to as the user's message. A piece of footprint information set by the user may include one or more of text, picture, audio, video, and three-dimensional model at the same time. By enabling the user's footprint information to include at least one of text, pictures, audio, video, and a three-dimensional model, it is beneficial to enrich the expression form of the user's footprint information, thereby helping to enrich the way for the user to interact with the three-dimensional space scene.
在一个可选示例中,获取用户在三维空间场景中的当前视角所对应的当前视图中的第一像素点可以为:获取用户在三维空间场景中的当前视角所对应的当前视图的中心像素点,并将该中心像素点作为第一像素点。例如,假设用户在三维空间场景中的当前视角,通过点击按钮或者菜单上的选项等方式触发了在三维空间场景中设置足迹信息的功能,此时,可以直接将当前视图的中心像素点作为第一像素点。中心像素点可以认为是针对用户的足迹信息设置的缺省像素点,用户可以通过拖拽等方式对该缺省像素点进行更改。在一个例子中,中心像素点可以认为是当前视图的中心区域中的一个像素点。当前视图的中心区域可以包括一个像素点或者多个像素点。通过直接将当前视图的中心像素点作为第一像素点,不仅有利于快速获得第一像素点,而且,有利于使用户设置的足迹信息在当前视图中位于较为显著的位置。In an optional example, obtaining the first pixel point in the current view corresponding to the current perspective of the user in the three-dimensional space scene may be: obtaining the central pixel point of the current view corresponding to the current perspective of the user in the three-dimensional space scene , And use the central pixel as the first pixel. For example, suppose that the user’s current perspective in the 3D space scene triggers the function of setting footprint information in the 3D space scene by clicking a button or an option on the menu. In this case, you can directly use the center pixel of the current view as the first One pixel. The center pixel can be considered as the default pixel set for the user's footprint information, and the user can change the default pixel by dragging and other methods. In an example, the central pixel can be considered as a pixel in the central area of the current view. The central area of the current view may include one pixel or multiple pixels. By directly using the central pixel of the current view as the first pixel, it is not only beneficial to quickly obtain the first pixel, but also beneficial to make the footprint information set by the user be located in a more prominent position in the current view.
在一个可选示例中,获取用户在三维空间场景中的当前视角所对应的当前视图中的第一像素点可以为:根据用户在三维空间场景中的当前视角所对应的当前视图中设置足迹信息目标位置的操作,获取足迹信息目标位置对应的当前视图中的像素点,该像素点被作为第一像素点。也就是说,在用户执行设置足迹信息目标位置的操作的情况下,可以将该操作在当前视图中形成的足迹信息目标位置所在的像素点,作为第一像素点。In an optional example, obtaining the first pixel in the current view corresponding to the current perspective of the user in the three-dimensional space scene may be: setting the footprint information in the current view corresponding to the current perspective of the user in the three-dimensional space scene The operation of the target position obtains the pixel point in the current view corresponding to the target position of the footprint information, and the pixel point is regarded as the first pixel point. That is, when the user performs the operation of setting the target position of the footprint information, the pixel point where the target position of the footprint information formed by the operation in the current view is located may be used as the first pixel point.
可选地,设置足迹信息目标位置的操作可以为用于确定足迹信息的起始目标位置的操作,也可以为用于确定足迹信息的终止目标位置的操作,还可以为用于确定足迹信息的中心目标位置的操作。Optionally, the operation of setting the target position of the footprint information may be an operation used to determine the starting target position of the footprint information, an operation used to determine the ending target position of the footprint information, or an operation used to determine the footprint information. Operation of the center target position.
可选地,设置足迹信息目标位置的操作可以具体为基于鼠标或者键盘等工具的点击操作或者滚动操作或者拖动操作等,也可以具体为基于触摸屏的点击操作或者拖拽操作等。本公开对设置足迹信息目标位置的具体操作不作限定。Optionally, the operation of setting the target location of the footprint information may specifically be a click operation or a scroll operation or drag operation based on a tool such as a mouse or a keyboard, and may also be a click operation or a drag operation based on a touch screen. The present disclosure does not limit the specific operation of setting the target position of the footprint information.
通过根据用户的设置足迹信息目标位置的操作来确定第一像素点,有利于使用户设置的足迹信息位于用户期望的位置,从而有利于提高设置足迹信息的灵活性,并有利于使足迹信息的位置更为恰当。By determining the first pixel point according to the user's operation of setting the target position of the footprint information, it is beneficial to make the footprint information set by the user be located at the desired position of the user, thereby improving the flexibility of setting the footprint information and making the footprint information more stable. The location is more appropriate.
可选地,假设在用户基于其在三维空间场景中的当前视角观看当前视图浏览的过程中,用户通过点击按钮或者菜单上的选项等方式触发了在三维空间场景中设置足迹信息的功能,此时,用户可以利用鼠标左键点击、键盘中的上下左右键移动光标或者点击触摸屏中的相应位置 等方式,来设置其所希望的足迹信息在当前视图中的位置。可以将该位置所在的像素点作为第一像素点。Optionally, suppose that while the user is viewing the current view based on his current perspective in the 3D space scene, the user triggers the function of setting footprint information in the 3D space scene by clicking a button or an option on the menu. At the time, the user can use the left mouse button to click, the keyboard up, down, left, and right keys to move the cursor or click the corresponding position on the touch screen to set the position of the desired footprint information in the current view. The pixel at this position can be used as the first pixel.
可选地,假设在用户基于其在三维空间场景中的当前视角观看当前视图浏览的过程中,用户通过点击按钮或者菜单上的选项等方式触发了在三维空间场景中设置足迹信息的功能,此时,可以先将当前视图的中心像素点作为第一像素点。如果用户不对该第一像素点进行更改,则将该中心像素点作为最终的第一像素点。如果用户利用鼠标左键的拖动操作、键盘中的上下左右键移动光标操作或者用手指在触摸屏上的拖动操作等方式,来更改第一像素点,则将操作结果获得的具体位置所在的像素点作为第一像素点。Optionally, suppose that while the user is viewing the current view based on his current perspective in the 3D space scene, the user triggers the function of setting footprint information in the 3D space scene by clicking a button or an option on the menu. When, you can first use the center pixel of the current view as the first pixel. If the user does not modify the first pixel, then the center pixel is regarded as the final first pixel. If the user uses the left mouse button drag operation, the keyboard up, down, left, and right buttons to move the cursor, or the finger drag operation on the touch screen to change the first pixel, the specific position obtained by the operation result will be changed. The pixel point is regarded as the first pixel point.
在一个可选示例中,确定第一像素点对应的三维模型(步骤220)的实现方式可以如图3所示。如图3所示,步骤220还包括步骤310至340。In an optional example, the implementation of determining the three-dimensional model corresponding to the first pixel (step 220) may be as shown in FIG. 3. As shown in FIG. 3, step 220 further includes steps 310 to 340.
在步骤310,确定点前视图的中心像素点作为第一像素点。In step 310, the central pixel point of the front view is determined as the first pixel point.
可选地,中心像素点可以认为是针对用户的足迹信息设置的缺省像素点。在一个例子中,假设当前视图为(2n+1)×(2m+1)的图像(其中的n和m均为大于1的整数),则可以直接将当前视图中的像素点(n+1,m+1)作为中心像素点。在另一个例子中,假设当前视图为2n×2m的图像(其中的n和m均为大于1的整数),则可以将当前视图中的像素点(n,m)、像素点(n+1,m)、像素点(n,m+1)以及像素点(n+1,m+1)作为当前视图的中心区域,从而可以将中心区域中的其中任一像素点作为中心像素点。Optionally, the central pixel may be considered as the default pixel set for the user's footprint information. In an example, assuming that the current view is an image of (2n+1)×(2m+1) (where n and m are both integers greater than 1), then the pixels in the current view (n+1 , M+1) as the central pixel. In another example, assuming that the current view is an image of 2n×2m (where n and m are both integers greater than 1), the pixel point (n, m) and pixel point (n+1) in the current view can be , M), pixel (n, m+1), and pixel (n+1, m+1) are used as the central area of the current view, so that any pixel in the central area can be used as the central pixel.
在步骤320,确定针对第一像素点是否设置有三维模型。如果针对第一像素点设置有三维模型,则到步骤330。如果针对第一像素点未设置有三维模型,则到步骤340。In step 320, it is determined whether a three-dimensional model is set for the first pixel. If a three-dimensional model is set for the first pixel point, go to step 330. If the three-dimensional model is not set for the first pixel point, go to step 340.
在图3的示例中,由于并不是当前视图中的所有像素点均为三维模型中的相应点的呈现,因此需要判断针对第一像素点是否设置有三维模型,即需要判断第一像素点是否为用于呈现三维模型中的相应点的像素点,以使得可以将用户的足迹信息设置在三维模型中的相应位置处。In the example in Figure 3, since not all pixels in the current view are the representation of corresponding points in the three-dimensional model, it is necessary to determine whether a three-dimensional model is set for the first pixel, that is, it is necessary to determine whether the first pixel is It is the pixel point used to present the corresponding point in the three-dimensional model, so that the user's footprint information can be set at the corresponding position in the three-dimensional model.
在步骤330,响应于针对第一像素点设置有三维模型的确定,将针对第一像素点设置的三维模型作为第一像素点对应的三维模型。In step 330, in response to the determination that the three-dimensional model is set for the first pixel, the three-dimensional model set for the first pixel is used as the three-dimensional model corresponding to the first pixel.
在步骤340,响应于针对第一像素点未设置有三维模型的确定,将针对当前视图中的其他像素点设置的三维模型作为第一像素点对应的三维模型。In step 340, in response to the determination that the three-dimensional model is not set for the first pixel, the three-dimensional model set for other pixels in the current view is used as the three-dimensional model corresponding to the first pixel.
可选地,当前视图中的其他像素点为当前视图中的设置有三维模型的像素点。可以根据预设规则来寻找设置有三维模型的像素点。在一个例子中,寻找到的其他像素点可以为在某一个方向(如左方向、右方向、上方向或者下方向)上距离第一像素点最近的像素点。Optionally, other pixels in the current view are pixels in the current view where the three-dimensional model is set. The pixels where the three-dimensional model is set can be found according to preset rules. In an example, the other pixels found may be the pixels closest to the first pixel in a certain direction (such as the left direction, the right direction, the upper direction, or the lower direction).
可选地,可以以第一像素点为起点,根据预设检查规则,对三维空间场景中的当前视角所 对应的当前视图中的像素点进行检查,如果确定检查到设置有三维模型的像素点,则获得第一像素点对应的三维模型,并停止本次检查过程。例如,可以以第一像素点为起点,向左检查当前视图中的像素点,并判断是否针对当前检查到的像素点设置了三维模型。如果判断结果为针对当前检查到的像素点设置有三维模型,则停止本次检查过程,并将当前检查获得的三维模型作为第一像素点对应的三维模型。另外,可以利用检查到的设置有三维模型的像素点更新第一像素点。当然,也可以不更新第一像素点。Optionally, the first pixel point can be used as a starting point, and according to a preset inspection rule, the pixel point in the current view corresponding to the current perspective in the three-dimensional space scene can be checked. If it is determined that the pixel point with the three-dimensional model is checked , The three-dimensional model corresponding to the first pixel is obtained, and the inspection process is stopped. For example, you can use the first pixel as a starting point to check the pixels in the current view to the left, and determine whether a three-dimensional model is set for the currently checked pixel. If the result of the judgment is that a three-dimensional model is set for the currently inspected pixel, the inspection process is stopped, and the three-dimensional model obtained by the current inspection is used as the three-dimensional model corresponding to the first pixel. In addition, the first pixel point may be updated by using the detected pixel point provided with the three-dimensional model. Of course, the first pixel may not be updated.
通过判断第一像素点是否设置有三维模型,并根据判断结果执行不同的操作,有利于避免在针对第一像素点未设置有三维模型的情况下,存在的无法将用户的足迹信息设置在三维模型中的相应位置处的现象。进一步,通过利用预设检查规则,获得设置有三维模型的其他像素点,并将针对该其他像素点设置的三维模型作为第一像素点对应的三维模型,有利于快速获得第一像素点对应的三维模型。By judging whether the first pixel is provided with a three-dimensional model, and performing different operations according to the judgment result, it is helpful to avoid the inability to set the user’s footprint information in the three-dimensional when the first pixel is not provided with a three-dimensional model. The phenomenon at the corresponding position in the model. Further, by using preset inspection rules to obtain other pixels that are provided with a three-dimensional model, and use the three-dimensional model set for the other pixels as the three-dimensional model corresponding to the first pixel, it is beneficial to quickly obtain the corresponding three-dimensional model of the first pixel. Three-dimensional model.
在一个可选示例中,确定第一像素点对应的三维模型(步骤220)的实现方式可以如图4所示。如图4所示,步骤220可以包括步骤410至步骤450。In an optional example, the implementation of determining the three-dimensional model corresponding to the first pixel (step 220) may be as shown in FIG. 4. As shown in FIG. 4, step 220 may include step 410 to step 450.
在步骤410,响应于用户在当前视图中设置足迹信息的目标位置的操作,确定足迹信息的目标位置对应的当前视图中的像素点,以作为第一像素点。In step 410, in response to the user's operation of setting the target position of the footprint information in the current view, a pixel point in the current view corresponding to the target position of the footprint information is determined as the first pixel point.
可选地,可以允许用户在当前视图中自行设置足迹信息的具体位置(即足迹信息的目标位置)。例如,在用户触发在三维空间场景中设置足迹信息的功能后,用户可以通过在触摸屏上点击、滑动、拖拽等操作,在当前视图中设置足迹信息的目标位置。该足迹信息的目标位置可以为文本框的左上顶点、左下顶点、右上顶点或者右下顶点等。该足迹信息的目标位置可以为图片的左上顶点、左下顶点、右上顶点或者右下顶点等。根据本公开的示范性实施例,足迹信息的目标位置可以是当前视图中的一个像素点,该像素点即为第一像素点。Optionally, the user may be allowed to set the specific location of the footprint information (that is, the target location of the footprint information) in the current view. For example, after the user triggers the function of setting the footprint information in the three-dimensional space scene, the user can set the target position of the footprint information in the current view by tapping, sliding, dragging and other operations on the touch screen. The target position of the footprint information may be the upper left vertex, the lower left vertex, the upper right vertex, or the lower right vertex of the text box. The target location of the footprint information may be the upper left vertex, the lower left vertex, the upper right vertex, or the lower right vertex of the picture. According to an exemplary embodiment of the present disclosure, the target location of the footprint information may be a pixel point in the current view, and the pixel point is the first pixel point.
在步骤420,判断针对第一像素点是否设置有三维模型。如果针对第一像素点设置有三维模型,则到步骤430。如果针对第一像素点未设置有三维模型,则到步骤440。In step 420, it is determined whether a three-dimensional model is set for the first pixel. If a three-dimensional model is set for the first pixel point, go to step 430. If a three-dimensional model is not set for the first pixel point, go to step 440.
在步骤430,响应于针对第一像素点设置有三维模型的确定,将针对第一像素点设置的三维模型作为第一像素点对应的三维模型。In step 430, in response to the determination that the three-dimensional model is set for the first pixel, the three-dimensional model set for the first pixel is used as the three-dimensional model corresponding to the first pixel.
在步骤440,响应于针对第一像素点未设置有三维模型的确定,输出更新足迹信息的目标位置的提示信息。In step 440, in response to the determination that the three-dimensional model is not set for the first pixel point, prompt information for updating the target position of the footprint information is output.
可选地,提示信息用于提示用户更新其当前设置的足迹信息目标位置。即,提示信息用于提示用户其当前设置的足迹信息目标位置不能设置足迹信息,用户应重新设置足迹信息目标位置。可以通过文字或者音频或者图形等形式输出提示信息。在输出提示信息后,等待用户的后 续操作,如果用户此时触发了取消设置足迹信息的功能,则图4所示的流程结束。Optionally, the prompt information is used to prompt the user to update the target location of the footprint information currently set. That is, the prompt information is used to prompt the user that the current target location of the footprint information cannot be set to the footprint information, and the user should reset the target location of the footprint information. The prompt information can be output in the form of text, audio, or graphics. After outputting the prompt information, wait for the user's subsequent operations. If the user triggers the function of canceling the setting of the footprint information at this time, the process shown in Figure 4 ends.
在步骤450,响应于针对更新后的足迹信息的目标位置对应的当前视图中的像素点设置有三维模型的确定,将该设置有三维模型的像素点作为第一像素点。然后,流程返回步骤420。In step 450, in response to the determination that the pixel in the current view corresponding to the target position of the updated footprint information is set with a three-dimensional model, the pixel set with the three-dimensional model is taken as the first pixel. Then, the flow returns to step 420.
如果用户当前执行了更新足迹信息目标位置的操作,则再次获得足迹信息目标位置。再次获得的足迹信息目标位置同样可以是当前视图中的一个像素点,该像素点即为第一像素点。也就是说,前次获得的第一像素点被当前获得的足迹信息目标位置所在的相似度更新。If the user currently performs the operation of updating the target position of the footprint information, the target position of the footprint information is obtained again. The target location of the footprint information obtained again may also be a pixel in the current view, and this pixel is the first pixel. That is, the first pixel point obtained last time is updated by the similarity of the target position of the currently obtained footprint information.
通过判断用户自行设置的足迹信息目标位置是否设置有三维模型,并根据判断结果执行不同的操作,有利于避免在针对足迹信息目标位置未设置有三维模型的情况下,存在的无法将用户的足迹信息设置在三维模型中的相应位置处的现象。通过利用步骤420至步骤450的循环过程,有利于促使用户将其足迹信息最终设置在三维模型的相应位置处,从而有利于使足迹信息的位置更为恰当。By judging whether the target location of the footprint information set by the user is provided with a three-dimensional model, and performing different operations according to the judgment result, it is helpful to avoid the inability to remove the user's footprint when the target location of the footprint information is not provided with a three-dimensional model. The phenomenon that information is set at the corresponding position in the three-dimensional model. By using the cyclic process from step 420 to step 450, it is helpful to prompt the user to finally set his footprint information at the corresponding position of the three-dimensional model, thereby helping to make the position of the footprint information more appropriate.
在一个可选示例中,在第一像素点设置有三维模型的情况下,由于当前视图中的第一像素点与三维模型中的点存在映射关系,因此,可以基于该映射关系获得第一像素点所对应的三维模型中的点,该点的位置即为第一像素点在三维模型中的位置。可以直接将第一像素点在三维模型中的位置作为用户的足迹信息在三维模型中的位置,从而有利于快捷准确地获得用户的足迹信息在三维模型中的位置。In an optional example, when the first pixel is provided with a three-dimensional model, since the first pixel in the current view has a mapping relationship with the point in the three-dimensional model, the first pixel can be obtained based on the mapping relationship. The point corresponding to the point in the three-dimensional model, the position of the point is the position of the first pixel point in the three-dimensional model. The position of the first pixel in the three-dimensional model can be directly used as the position of the user's footprint information in the three-dimensional model, which is beneficial to quickly and accurately obtain the position of the user's footprint information in the three-dimensional model.
在一个可选示例中,在浏览用户观看三维空间场景的过程中,可以为该浏览用户呈现至少一用户在该三维空间场景中留下的足迹信息。一个例子如图5所示。In an optional example, in the process of viewing the three-dimensional space scene by the browsing user, the browsing user may be presented with at least one user's footprint information left in the three-dimensional space scene. An example is shown in Figure 5.
图5中,在步骤510,对于浏览三维空间场景的任一浏览用户,确定该浏览用户在三维空间场景中的当前视角所对应的足迹区域。In FIG. 5, in step 510, for any browsing user browsing the three-dimensional space scene, the footprint area corresponding to the current perspective of the browsing user in the three-dimensional space scene is determined.
可选地,浏览用户包括在该三维空间场景中设置其足迹信息的用户。足迹区域可以认为是针对需要显示的足迹信息设置的区域。该足迹区域可以为基于当前视图的足迹区域,也可以为基于三维模型的足迹区域。足迹区域的大小可以是预先设置的。足迹区域的形状可以为长方形或者圆形或者三角形等。Optionally, the browsing user includes a user who sets his footprint information in the three-dimensional space scene. The footprint area can be considered as an area set for the footprint information that needs to be displayed. The footprint area can be a footprint area based on the current view, or a footprint area based on a three-dimensional model. The size of the footprint area can be preset. The shape of the footprint area can be rectangle, circle, triangle, etc.
在足迹区域为基于当前视图的足迹区域时,根据本公开的一些示范性实施例,确定足迹区域的一种实现方式可以为:首先,获取浏览用户在三维空间场景中的当前视角所对应的当前视图的中心像素点,之后,以该中心像素点为圆心,以预定长度(如三维空间场景中的1.5米等,且1.5米可以被换算为当前视图中的长度)为半径,确定当前视图中的足迹区域。由于当前视图中的足迹区域中的至少部分像素点与三维模型中的点存在映射关系,因此,利用当前视图中的足迹区域,可以便捷的获得当前需要显示的足迹信息。另外,当前视图中的足迹区域可以认 为是一个圆形,即当前视图中的足迹区域不具有深度信息。When the footprint area is a footprint area based on the current view, according to some exemplary embodiments of the present disclosure, an implementation manner of determining the footprint area may be: First, obtain the current view corresponding to the current viewing angle of the browsing user in the three-dimensional space scene. The center pixel of the view, and then the center pixel is the center of the circle, and the predetermined length (such as 1.5 meters in a three-dimensional scene, etc., and 1.5 meters can be converted to the length in the current view) is used as the radius to determine the current view Footprint area. Since at least some of the pixels in the footprint area in the current view have a mapping relationship with points in the three-dimensional model, the footprint information that currently needs to be displayed can be easily obtained by using the footprint area in the current view. In addition, the footprint area in the current view can be regarded as a circle, that is, the footprint area in the current view does not have depth information.
在足迹区域为基于三维模型的足迹区域时,根据本公开的一些示范性实施例,确定足迹区域的一种实现方式可以为:首先,获取浏览用户在三维空间场景中的当前视角所对应的当前视图的中心像素点,并确定该中心像素点是否设置有三维模型,如果该中心像素点设置有三维模型,则确定该中心像素点在该三维模型中的位置,之后,以该位置为圆心,以预定长度(如三维空间场景中的1.5米等)为半径,确定三维模型中的足迹区域。该足迹区域可能会完全位于一个三维模型中,也可能会跨多个三维模型。另外,三维模型中的足迹区域可以认为是一个圆柱体,即三维模型中的足迹区域具有深度信息。When the footprint area is a footprint area based on a three-dimensional model, according to some exemplary embodiments of the present disclosure, an implementation manner for determining the footprint area may be: First, obtain the current view corresponding to the current perspective of the browsing user in the three-dimensional space scene. View the center pixel of the view, and determine whether the center pixel is set with a three-dimensional model. If the center pixel is set with a three-dimensional model, determine the position of the center pixel in the three-dimensional model, and then use that position as the center of the circle, Use a predetermined length (such as 1.5 meters in a three-dimensional space scene, etc.) as a radius to determine the footprint area in the three-dimensional model. The footprint area may be completely in one 3D model, or it may span multiple 3D models. In addition, the footprint area in the three-dimensional model can be considered as a cylinder, that is, the footprint area in the three-dimensional model has depth information.
在步骤520,确定三维模型中的属于足迹区域的足迹信息。In step 520, the footprint information belonging to the footprint area in the three-dimensional model is determined.
在足迹区域为基于当前视图的足迹区域时,本公开的实施例可以检查足迹区域中的每一个像素点是否与三维模型中的点存在映射关系。如果存在映射关系,再判断与像素点存在映射关系的三维模型中的点是否设置有足迹信息。如果设置有足迹信息,则可以将该足迹信息认为是属于该足迹区域的足迹信息。When the footprint area is a footprint area based on the current view, the embodiment of the present disclosure can check whether each pixel point in the footprint area has a mapping relationship with a point in the three-dimensional model. If there is a mapping relationship, then it is determined whether the points in the three-dimensional model that have a mapping relationship with the pixel points are provided with footprint information. If the footprint information is set, the footprint information can be regarded as the footprint information belonging to the footprint area.
在足迹区域为基于三维模型的足迹区域时,本公开的实施例可以检查足迹区域中的每一个点是否设置有足迹信息。如果设置有足迹信息,则可以将该足迹信息认为是属于该足迹区域的足迹信息。When the footprint area is a footprint area based on a three-dimensional model, the embodiments of the present disclosure can check whether each point in the footprint area is provided with footprint information. If the footprint information is set, the footprint information can be regarded as the footprint information belonging to the footprint area.
在步骤530,在该浏览用户在三维空间场景中的当前视角所对应的当前视图中,显示属于该足迹区域的足迹信息。In step 530, in the current view corresponding to the current perspective of the browsing user in the three-dimensional space scene, the footprint information belonging to the footprint area is displayed.
可选地,可以根据各足迹信息的位置,确定出属于该足迹区域的各足迹信息分别在当前视图中的位置,从而可以根据各足迹信息分别在当前视图中的位置显示各足迹信息。在显示足迹信息的过程中,可以尽量避免不同足迹信息在当前视图中的重叠显示现象。Optionally, the location of each footprint information belonging to the footprint area in the current view can be determined according to the location of each footprint information, so that each footprint information can be displayed according to the location of each footprint information in the current view. In the process of displaying footprint information, it is possible to avoid overlapping display of different footprint information in the current view.
可选地,获得的多个足迹信息可能具有不同的位置,也可能具有相同的位置(即足迹信息的位置冲突)。响应于确定所述属于所述足迹区域的足迹信息具有不同位置的多个足迹信息,可以直接根据多个足迹信息分别在当前视图中的图像位置,在当前视图中显示各足迹信息。而且,可以允许显示的各足迹信息部分重叠,也可以通过位置控制使各足迹信息互不重叠。响应于确定所述属于所述足迹区域的足迹信息包括相同位置的不同足迹信息,可以在当前视图中为不同足迹信息分别分配不同的图像位置,并根据分配的图像位置,在当前视图中显示上述具有相同位置的不同足迹信息,从而有利于避免不同足迹信息在当前视图中的重叠显示现象。Optionally, the obtained multiple footprint information may have different positions, or may have the same position (that is, the position of the footprint information conflicts). In response to determining that the footprint information belonging to the footprint area has multiple footprint information with different positions, each footprint information may be displayed in the current view directly according to the image positions of the multiple footprint information in the current view. Moreover, the displayed footprint information can be allowed to partially overlap, and the location control can also be used to make the footprint information not overlap each other. In response to determining that the footprint information belonging to the footprint area includes different footprint information at the same location, different image positions may be assigned to different footprint information in the current view, and the image positions described above may be displayed in the current view according to the assigned image positions. Having different footprint information at the same position helps to avoid overlapping display of different footprint information in the current view.
可选地,可以显示属于该足迹区域的所有足迹信息,也可以显示属于该足迹区域的部分足迹信息。例如,在属于该足迹区域的所有足迹信息的数量过于庞大(例如,数量超过预定数量) 时,可以按照预定规则,从中选取部分足迹信息,并在当前视图中显示选取出的部分足迹信息。Optionally, all the footprint information belonging to the footprint area can be displayed, or part of the footprint information belonging to the footprint area can be displayed. For example, when the amount of all the footprint information belonging to the footprint area is too large (for example, the number exceeds a predetermined number), part of the footprint information can be selected from it according to a predetermined rule, and the selected part of the footprint information can be displayed in the current view.
可选地,可以从属于该足迹区域的所有足迹信息中随机选取出预定数量的足迹信息,并在当前视图中显示随机选取出的部分足迹信息。Optionally, a predetermined number of footprint information can be randomly selected from all the footprint information belonging to the footprint area, and part of the randomly selected footprint information can be displayed in the current view.
可选地,可以从属于该足迹区域的所有足迹信息中优先选取浏览用户自己设置的足迹信息,还可以优先选取质量好的足迹信息等,并在当前视图中显示选取出的部分足迹信息。Optionally, it is possible to preferentially select and browse the footprint information set by the user from all the footprint information belonging to the footprint area, or to preferentially select good-quality footprint information, etc., and display part of the selected footprint information in the current view.
在一个可选示例中,可以采用弹幕的形式,为浏览用户显示当前视图之外的足迹信息。例如,可以先确定三维模型中的不属于当前视图的所有足迹信息,并以弹幕的形式,在该浏览用户在三维空间场景中的当前视角所对应的当前视图中,显示上述所有足迹信息或者部分足迹信息。In an optional example, the form of a bullet screen may be used to display the footprint information outside the current view for the browsing user. For example, you can first determine all the footprint information in the three-dimensional model that does not belong to the current view, and display all the above-mentioned footprint information in the current view corresponding to the current perspective of the browsing user in the three-dimensional scene in the form of a bullet screen. Part of the footprint information.
在一个可选示例中,可以采用弹幕的形式,为浏览用户显示足迹区域之外的足迹信息。例如,可以先确定三维模型中的不属于足迹区域的所有足迹信息,并以弹幕的形式,在该浏览用户在三维空间场景中的当前视角所对应的当前视图中,显示上述所有足迹信息或者部分足迹信息。In an optional example, the form of a bullet screen may be used to display the footprint information outside the footprint area for the browsing user. For example, you can first determine all the footprint information in the three-dimensional model that does not belong to the footprint area, and display all the above-mentioned footprint information in the current view corresponding to the current perspective of the browsing user in the three-dimensional scene in the form of a bullet screen. Part of the footprint information.
通过采用弹幕的形式,显示不属于足迹区域/当前视图的足迹信息,不仅有利于促使浏览用户探查该三维空间场景的其他部分,提高浏览用户的沉浸感,而且有利于进一步提升浏览用户的VR全景体验。By adopting the form of barrage to display the footprint information that does not belong to the footprint area/current view, it is not only conducive to prompting the browsing user to explore other parts of the three-dimensional space scene, improving the immersion of the browsing user, but also conducive to further improving the VR of the browsing user Panoramic experience.
图6为本公开的用于实现三维空间场景互动的装置一个实施例的结构示意图。该实施例的装置可用于实现本公开上述各方法实施例。FIG. 6 is a schematic structural diagram of an embodiment of an apparatus for realizing interaction in a three-dimensional space scene of the present disclosure. The device of this embodiment can be used to implement the foregoing method embodiments of the present disclosure.
如图6所示,本实施例的装置包括:获取像素点模块600、确定三维模型模块601、确定位置模块602以及设置足迹信息模块603。另外,装置还可以包括:确定足迹区域模块604、确定足迹信息模块605、显示足迹信息模块606以及弹幕显示模块607。As shown in FIG. 6, the device of this embodiment includes: a pixel point acquiring module 600, a three-dimensional model determining module 601, a position determining module 602, and a footprint information setting module 603. In addition, the device may further include: a footprint area determination module 604, a footprint information determination module 605, a footprint information display module 606, and a bullet screen display module 607.
获取像素点模块600用于响应于检测到在三维空间场景中设置足迹信息的用户操作,确定用户在三维空间场景中的当前视角所对应的当前视图中的第一像素点。The pixel obtaining module 600 is configured to determine the first pixel in the current view corresponding to the current perspective of the user in the three-dimensional space scene in response to detecting the user operation of setting the footprint information in the three-dimensional space scene.
可选地,足迹信息可以包括:文本、图片、音频、视频以及三维模型中的至少一个。Optionally, the footprint information may include: at least one of text, picture, audio, video, and a three-dimensional model.
可选地,获取像素点模块600可以包括:第一子模块6001。该第一子模块6001用于确定所述当前视图的中心像素点作为第一像素点。Optionally, the pixel point acquiring module 600 may include: a first sub-module 6001. The first sub-module 6001 is used to determine the center pixel of the current view as the first pixel.
可选地,获取像素点模块600可以包括:第五子模块6002。该第五子模块6002用于响应于用户在三维空间场景中的当前视角所对应的当前视图中设置足迹信息的目标位置的操作,确定足迹信息的目标位置对应的当前视图中的像素点。第五子模块6002可以将该像素点做为第一像素点。Optionally, the pixel point obtaining module 600 may include: a fifth sub-module 6002. The fifth sub-module 6002 is configured to determine the pixel points in the current view corresponding to the target position of the footprint information in response to the user's operation of setting the target position of the footprint information in the current view corresponding to the current perspective in the three-dimensional space scene. The fifth sub-module 6002 can use the pixel as the first pixel.
确定三维模型模块601用于确定获取像素点模块600获取到的第一像素点对应的三维模型。The three-dimensional model determining module 601 is used to determine the three-dimensional model corresponding to the first pixel obtained by the pixel obtaining module 600.
可选地,在获取像素点模块600包括第一子模块6001的情况下,确定三维模型模块601可以包括:第二子模块6011、第三子模块6012以及第四子模块6013。第二子模块6011用于确定针对第一像素点是否设置有三维模型。第三子模块6012用于如果第二子模块6011的确定结果为针对第一像素点设置有三维模型,则将针对第一像素点设置的三维模型作为第一像素点对应的三维模型。第四子模块6013用于如果第二子模块6011的判断结果为针对第一像素点未设置有三维模型,则将针对当前视图中的其他像素点设置的三维模型作为第一像素点对应的三维模型。例如,如果第二子模块6011的判断结果为针对第一像素点未设置有三维模型,则第四子模块6013可以以第一像素点为起点,根据预设检查规则,对三维空间场景中的当前视角所对应的当前视图中的其他像素点进行检查。如果检查到设置有三维模型的像素点,则将第一像素点更新为设置有三维模型的像素点,获得第一像素点对应的三维模型,并停止本次检查。Optionally, in a case where the pixel point obtaining module 600 includes the first sub-module 6001, the determining three-dimensional model module 601 may include: the second sub-module 6011, the third sub-module 6012, and the fourth sub-module 6013. The second sub-module 6011 is used to determine whether a three-dimensional model is set for the first pixel. The third sub-module 6012 is configured to, if the determination result of the second sub-module 6011 is that a three-dimensional model is set for the first pixel, use the three-dimensional model set for the first pixel as the three-dimensional model corresponding to the first pixel. The fourth sub-module 6013 is configured to, if the judgment result of the second sub-module 6011 is that no three-dimensional model is set for the first pixel, use the three-dimensional model set for other pixels in the current view as the three-dimensional model corresponding to the first pixel. Model. For example, if the judgment result of the second sub-module 6011 is that a three-dimensional model is not set for the first pixel, the fourth sub-module 6013 can take the first pixel as a starting point and perform a check on the three-dimensional scene according to the preset inspection rules. Check other pixels in the current view corresponding to the current angle of view. If a pixel with a three-dimensional model is detected, the first pixel is updated to a pixel with a three-dimensional model, the three-dimensional model corresponding to the first pixel is obtained, and this inspection is stopped.
在获取像素点模块600包括第五子模块6002的情况下,确定三维模型模块601可以包括:第六子模块6014、第七子模块6015以及第八子模块6016。第六子模块6014用于确定针对第一像素点是否设置有三维模型。如果第六子模块6014的确定结果为针对第一像素点设置有三维模型,则第七子模块6015将针对第一像素点设置的三维模型作为第一像素点对应的三维模型。如果第六子模块6014的确定结果为针对第一像素点未设置有三维模型,则第八子模块6016可以输出更新足迹信息目标位置的提示信息,并在第六子模块6014判断出针对更新后的足迹信息目标位置对应的当前视图中的像素点设置有三维模型时,将该设置有三维模型的像素点作为第一像素点。第八子模块6016获得第一像素点对应的三维模型。In the case where the pixel point obtaining module 600 includes a fifth sub-module 6002, the determining three-dimensional model module 601 may include: a sixth sub-module 6014, a seventh sub-module 6015, and an eighth sub-module 6016. The sixth sub-module 6014 is used to determine whether a three-dimensional model is set for the first pixel. If the determination result of the sixth sub-module 6014 is that a three-dimensional model is set for the first pixel, the seventh sub-module 6015 uses the three-dimensional model set for the first pixel as the three-dimensional model corresponding to the first pixel. If the determination result of the sixth sub-module 6014 is that no three-dimensional model is set for the first pixel, the eighth sub-module 6016 may output prompt information for updating the target position of the footprint information, and the sixth sub-module 6014 determines When the pixel in the current view corresponding to the target location of the footprint information is set with a three-dimensional model, the pixel with the three-dimensional model is used as the first pixel. The eighth sub-module 6016 obtains the three-dimensional model corresponding to the first pixel.
确定位置模块602用于确定用户的足迹信息在确定三维模型模块601确定出的三维模型中的位置。例如,确定位置模块602可以获取第一像素点在三维模型中的位置,且确定位置模块602可以将第一像素点在三维模型中的位置作为用户的足迹信息在三维模型中的位置。The position determining module 602 is used to determine the position of the user's footprint information in the three-dimensional model determined by the three-dimensional model determining module 601. For example, the position determining module 602 may obtain the position of the first pixel in the three-dimensional model, and the position determining module 602 may use the position of the first pixel in the three-dimensional model as the position of the user's footprint information in the three-dimensional model.
设置足迹信息模块603用于在确定位置模块602确定出的位置处设置用户的足迹信息。设置足迹信息模块603设置的用户的足迹信息用于显示给三维空间场景的浏览用户。The setting footprint information module 603 is used for setting the user's footprint information at the location determined by the location determining module 602. The user's footprint information set by the setting footprint information module 603 is used to be displayed to users who browse the three-dimensional space scene.
确定足迹区域模块604用于对于浏览三维空间场景的任一浏览用户,确定该浏览用户在三维空间场景中的当前视角所对应的足迹区域。例如,确定足迹区域模块604可以先确定该浏览用户在三维空间场景中的当前视角所对应的当前视图的中心像素点,然后,确定足迹区域模块604以该中心像素点为圆心,以预定长度为半径,确定当前视图中的足迹区域。The footprint area determination module 604 is used for determining the footprint area corresponding to the current perspective of the browsing user in the three-dimensional space scene for any browsing user who browses the three-dimensional space scene. For example, the module 604 for determining the footprint area may first determine the center pixel of the current view corresponding to the current perspective of the browsing user in the three-dimensional space scene, and then the module 604 for determining the footprint area 604 takes the center pixel as the center of the circle and the predetermined length as The radius determines the footprint area in the current view.
确定足迹信息模块605用于确定三维模型中的属于确定足迹区域模块604确定出的足迹 区域的足迹信息。The footprint information determining module 605 is used to determine the footprint information belonging to the footprint area determined by the footprint area determining module 604 in the three-dimensional model.
显示足迹信息模块606用于在该浏览用户在三维空间场景中的当前视角所对应的当前视图中,显示确定足迹信息模块605确定出的属于足迹区域的足迹信息。The footprint information display module 606 is configured to display the footprint information that belongs to the footprint area determined by the determination footprint information module 605 in the current view corresponding to the current perspective of the browsing user in the three-dimensional space scene.
可选地,响应于确定所述属于所述足迹区域的足迹信息具有不同位置的多个足迹信息,显示足迹信息模块606可以根据多个足迹信息分别在当前视图中的图像位置,在当前视图中显示所述多个足迹信息。Optionally, in response to determining that the footprint information belonging to the footprint area has a plurality of footprint information with different positions, the display footprint information module 606 may respectively locate the multiple footprint information in the current view according to the image positions of the multiple footprint information in the current view. Display the multiple footprint information.
可选地,响应于确定所述属于所述足迹区域的足迹信息具有相同位置的不同足迹信息,显示足迹信息模块606可以在当前视图中为不同足迹信息分配不同的图像位置,并根据分配的图像位置,在当前视图中显示不同足迹信息。Optionally, in response to determining that the footprint information belonging to the footprint area has different footprint information with the same position, the footprint information display module 606 may assign different image positions for different footprint information in the current view, and according to the assigned image Position, display different footstep information in the current view.
弹幕显示模块607用于确定三维模型中的不属于足迹区域/当前视图的至少一个足迹信息。弹幕显示模块607以弹幕的形式,在该浏览用户在三维空间场景中的当前视角所对应的当前视图中,显示该至少一个足迹信息。The barrage display module 607 is used to determine at least one piece of footprint information in the three-dimensional model that does not belong to the footprint area/current view. The barrage display module 607 displays the at least one footprint information in the current view corresponding to the current perspective of the browsing user in the three-dimensional space scene in the form of a barrage.
上述各模块及其包括的子模块具体执行的操作可以参见上述方法实施例中针对图2-图5的描述,在此不再详细说明。For the specific operations performed by the foregoing modules and the sub-modules included therein, reference may be made to the description of FIGS. 2 to 5 in the foregoing method embodiment, and detailed descriptions are omitted here.
请参考图7,示出了根据本公开的第一个基于三维模型的交互方法的一个实施例的流程700。该基于三维模型的交互方法应用于第一用户终端,第一用户终端呈现有用户界面,该基于三维模型的交互方法包括:Please refer to FIG. 7, which shows a process 700 of an embodiment of the first three-dimensional model-based interaction method according to the present disclosure. The three-dimensional model-based interaction method is applied to a first user terminal, and the first user terminal is presented with a user interface, and the three-dimensional model-based interaction method includes:
步骤710,响应于检测到用户针对用户界面的目标交互操作,向为用户界面提供页面数据的服务器发送针对目标交互操作的交互请求,其中,用户界面用于呈现三维模型,三维模型与第二用户终端登录的用户账号建立关联关系。Step 710: In response to detecting the user's target interaction operation on the user interface, send an interaction request for the target interaction operation to the server that provides page data for the user interface, where the user interface is used to present the three-dimensional model, the three-dimensional model and the second user The user account logged in by the terminal establishes an association relationship.
在本实施例中,用户可以使用第一用户终端通过网络与服务器进行交互。第一用户终端可以是各种电子设备,包括但不限于智能手机、平板电脑、膝上型便携计算机和台式计算机等等。第一用户终端可以安装有各种客户端应用,例如房产交易软件等。上述用户界面可以是第一用户终端所安装的应用中的页面。实践中,用户可以通过该用户界面与服务器进行交互,进而实现与其他用户终端(例如第二用户终端)之间的交互。In this embodiment, the user can use the first user terminal to interact with the server through the network. The first user terminal may be various electronic devices, including but not limited to smart phones, tablet computers, laptop portable computers, desktop computers, and so on. The first user terminal may be installed with various client applications, such as real estate transaction software. The aforementioned user interface may be a page in an application installed by the first user terminal. In practice, the user can interact with the server through the user interface, thereby realizing interaction with other user terminals (for example, the second user terminal).
在本实施例中,在检测到用户针对用户界面的目标交互操作的情况下,第一用户终端可以向为用户界面提供页面数据的服务器发送针对目标交互操作的交互请求。In this embodiment, in the case of detecting a user's target interaction operation on the user interface, the first user terminal may send an interaction request for the target interaction operation to a server that provides page data for the user interface.
上述用户界面用于呈现三维模型。三维模型与第二用户终端登录的用户账号预先建立关联关系。上述目标交互操作可以是各种用于指示第一用户终端请求与第二用户终端进行交互(信息交互)的操作。作为示例,该目标交互操作可以指示与第二用户终端进行视频通信。上述交 互请求可以用于指示第一用户终端的用户请求与第二用户终端进行交互。示例性的,上述交互请求可以用于指示第一用户终端的用户请求与第二用户终端进行视频通信。The aforementioned user interface is used to present a three-dimensional model. The three-dimensional model establishes an association relationship with the user account logged in by the second user terminal in advance. The aforementioned target interaction operation may be various operations for instructing the first user terminal to request interaction (information interaction) with the second user terminal. As an example, the target interaction operation may indicate video communication with the second user terminal. The foregoing interaction request may be used to indicate a user request of the first user terminal to interact with the second user terminal. Exemplarily, the foregoing interaction request may be used to instruct the user of the first user terminal to request video communication with the second user terminal.
在这里,在执行上述步骤710时,第一用户终端的用户界面可以呈现有上述三维模型,也可以未呈现三维模型。Here, when performing step 710, the user interface of the first user terminal may present the above-mentioned three-dimensional model, or may not present the three-dimensional model.
实践中,对于每个三维模型,其可以与一个用户账号预先建立有关联关系。由此,对于一个特定的三维模型,可以通过确定与该三维模型建立有关联关系的用户账号,从而确定登录该用户账号的用户终端,进而确定出用于与第一用户终端进行交互的用户终端(即第二用户终端)。In practice, for each three-dimensional model, it can be associated with a user account in advance. Therefore, for a specific three-dimensional model, the user account that is associated with the three-dimensional model can be determined to determine the user terminal logging in to the user account, and then the user terminal used to interact with the first user terminal is determined (Ie the second user terminal).
上述三维模型可以是任意物体的三维模型。示例性的,该三维模型可以是细胞内部的三维模型,也可以是房屋室内三维模型。The above-mentioned three-dimensional model may be a three-dimensional model of any object. Exemplarily, the three-dimensional model may be a three-dimensional model inside a cell, or a three-dimensional model of a house interior.
步骤720,接收服务器从第二用户终端获取的流媒体视频。Step 720: Receive the streaming video obtained by the server from the second user terminal.
在本实施例中,上述第一用户终端可以接收服务器从第二用户终端获取的流媒体视频。In this embodiment, the above-mentioned first user terminal may receive the streaming video obtained by the server from the second user terminal.
其中,上述交互确认信息可以用于指示所述第二用户终端的用户确认(同意)与第一用户终端进行上述交互请求指示的交互。例如,上述交互确认信息可以用于指示所述第二用户终端的用户确认(同意)与第一用户终端进行视频通信。The aforementioned interaction confirmation information may be used to instruct the user of the second user terminal to confirm (agree) to perform the interaction indicated by the aforementioned interaction request with the first user terminal. For example, the foregoing interactive confirmation information may be used to instruct the user of the second user terminal to confirm (agree) to conduct video communication with the first user terminal.
上述流媒体视频可以包含图像和/或语音。实践中,第二用户终端的图像获取装置和/或语音获取装置,可以用于获取上述流媒体视频。The aforementioned streaming video may include images and/or voice. In practice, the image acquisition device and/or the voice acquisition device of the second user terminal can be used to acquire the aforementioned streaming video.
实践中,服务器可以采用流媒体技术,将第二用户终端采集的图像和/或语音(即流媒体视频),持续发送至第一用户终端。其中,流媒体技术是指采用流式传输技术在网络上连续实时播放的媒体格式。流媒体技术也称流式媒体技术。这里,第二用户终端可以将其所采集的连续的影像和声音信息经过压缩处理后发送至服务器。由服务器向第一用户终端顺序或实时地传送各个压缩包,让使用第一用户终端的用户一边下载一边观看、收听。In practice, the server may use streaming media technology to continuously send the images and/or voice (ie streaming media video) collected by the second user terminal to the first user terminal. Among them, streaming media technology refers to a media format that uses streaming technology to be continuously played in real time on the network. Streaming media technology is also called streaming media technology. Here, the second user terminal may send the continuous image and sound information collected by it to the server after compression processing. The server transmits each compressed package to the first user terminal sequentially or in real time, so that users who use the first user terminal can watch and listen while downloading.
可选地,服务器可以将第二用户终端采集的流媒体视频发送至第一用户终端,也可以对第二用户终端采集的流媒体视频进行图像处理(例如美颜)、语音处理(例如去噪)、转码、录制、鉴黄等操作后,将处理后的流媒体视频发送至第一用户终端。Optionally, the server may send the streaming video collected by the second user terminal to the first user terminal, and may also perform image processing (such as beauty) and voice processing (such as denoising) on the streaming video collected by the second user terminal. ), after operations such as transcoding, recording, and pornography, send the processed streaming video to the first user terminal.
在本实施例的一些可选的实现方式中,可以在服务器接收到第二用户终端针对交互请求发送的交互确认信息的情况下,第一用户终端再执行上述步骤720。In some optional implementation manners of this embodiment, when the server receives the interaction confirmation information sent by the second user terminal in response to the interaction request, the first user terminal may perform step 720 again.
可以理解,上述可选的实现方式中,在第二用户终端针对交互请求发送的交互确认信息的情况下,通过后续步骤第一用户终端可以呈现流媒体视频;而在第二用户终端未发送上述交互确认信息的情况下,第一用户终端则不呈现流媒体视频。由此,可以在获得第二用户终端的用 户的允许(例如接通第一用户终端发起的视频通话)后,才在第一用户终端的用户界面上呈现流媒体视频和三维模型。这有助于提高对第二用户终端的用户的隐私保护性,为第二用户终端的用户向第一用户终端的用户呈现流媒体视频提供准备时间。It can be understood that, in the foregoing optional implementation manner, in the case of the interaction confirmation information sent by the second user terminal in response to the interaction request, the first user terminal can present the streaming video through subsequent steps; however, the second user terminal does not send the above In the case of interactive confirmation information, the first user terminal does not present the streaming video. Therefore, the streaming video and the three-dimensional model can be presented on the user interface of the first user terminal only after obtaining the permission of the user of the second user terminal (for example, connecting to the video call initiated by the first user terminal). This helps to improve the privacy protection of the user of the second user terminal, and provides preparation time for the user of the second user terminal to present the streaming media video to the user of the first user terminal.
在本实施例的一些可选的实现方式中,在服务器接收到交互请求之后,第一用户终端也可以直接执行上述步骤720(而无需第二用户终端针对交互请求发送的交互确认信息)。In some optional implementation manners of this embodiment, after the server receives the interaction request, the first user terminal may also directly execute the foregoing step 720 (without the interaction confirmation information sent by the second user terminal in response to the interaction request).
可以理解,上述可选的实现方式中,第二用户终端的用户可以处于向其他用户终端的用户拍摄流媒体视频(例如直播)的状态。由此,在服务器接收到交互请求之后,第一用户终端可以随时接收服务器从第二用户终端获取的流媒体视频,从而提高了流媒体视频呈现的实时性。It can be understood that, in the foregoing optional implementation manner, the user of the second user terminal may be in a state of shooting a streaming video (for example, a live broadcast) to users of other user terminals. Thus, after the server receives the interaction request, the first user terminal can receive the streaming video obtained by the server from the second user terminal at any time, thereby improving the real-time performance of the streaming video presentation.
在本实施例的一些可选的实现方式中,第一用户终端可以采用如下步骤接收服务器从第二用户终端获取的流媒体视频:In some optional implementation manners of this embodiment, the first user terminal may adopt the following steps to receive the streaming video obtained by the server from the second user terminal:
首先,将第一用户终端的当前网速值发送至服务器。First, the current network speed value of the first user terminal is sent to the server.
然后,接收服务器从所述第二用户终端获取并发送的的流媒体视频,该流媒体视频具有与当前网速值相匹配的分辨率。Then, the streaming media video obtained and sent by the server from the second user terminal is received, and the streaming media video has a resolution matching the current network speed value.
在这里,分辨率可以与网速值成正相关。Here, the resolution can be positively correlated with the network speed value.
可以理解,通过接收服务器发送的分辨率与当前网速值相匹配的、第二用户终端获取的流媒体视频,可以在网络较差的情况下,降低第一用户终端接收的流媒体视频的分辨率,以提高流媒体视频传输的实时性。It can be understood that the resolution of the streaming video received by the first user terminal can be reduced when the network is poor by receiving the streaming video sent by the server whose resolution matches the current network speed value and obtained by the second user terminal. Rate to improve the real-time performance of streaming video transmission.
步骤730,在用户界面上呈现流媒体视频和三维模型。Step 730: Present the streaming video and the three-dimensional model on the user interface.
在本实施例中,第一用户终端可以在用户界面上,同屏呈现流媒体视频和三维模型。In this embodiment, the first user terminal may present the streaming video and the three-dimensional model on the same screen on the user interface.
在这里,第一用户终端的上述用户界面可以被划分为两个部分,上述两个部分可以分别呈现流媒体视频和三维模型。可选地,也可以将三维模型作为上述用户界面的背景,在用户界面的一部分页面区域呈现流媒体视频。Here, the above-mentioned user interface of the first user terminal may be divided into two parts, and the above-mentioned two parts may respectively present a streaming video and a three-dimensional model. Optionally, the three-dimensional model can also be used as the background of the aforementioned user interface, and the streaming video is presented in a part of the page area of the user interface.
请参考图8A-8C,图8A-8C是针对图7的实施例的应用场景示意图。如图8A所示,在第一用户终端检测到用户针对用户界面的目标交互操作810(图示中目标交互操作810指示开启视频实时交互)的情况下,第一用户终端可以向为用户界面提供页面数据的服务器发送针对目标交互操作810的交互请求。其中,图8A中,用户界面呈现有XX家园的房屋室内的三维模型。该三维模型与第二用户终端登录的用户账号预先建立关联关系。在图8B中,第二用户终端接收到上述交互请求之后,第二用户终端的用户执行了开始交互的操作820。之后,第二用户终端向服务器发送针对交互请求的交互确认信息,以及第二用户终端采集的流媒体视频。最后,如图8C所示,第一用户终端在用户界面上呈现了流媒体视频830和三维模型。Please refer to FIGS. 8A-8C. FIGS. 8A-8C are schematic diagrams of application scenarios for the embodiment of FIG. 7. As shown in FIG. 8A, in the case where the first user terminal detects a user's target interaction operation 810 for the user interface (the target interaction operation 810 in the figure indicates to enable real-time video interaction), the first user terminal may provide the user interface with The server of the page data sends an interaction request for the target interaction operation 810. Among them, in FIG. 8A, the user interface presents a three-dimensional model of the house of XX home. The three-dimensional model has a pre-established association relationship with the user account logged in by the second user terminal. In FIG. 8B, after the second user terminal receives the aforementioned interaction request, the user of the second user terminal performs an operation 820 of starting interaction. After that, the second user terminal sends the interaction confirmation information for the interaction request and the streaming video collected by the second user terminal to the server. Finally, as shown in FIG. 8C, the first user terminal presents a streaming video 830 and a three-dimensional model on the user interface.
本公开的上述实施例提供的基于三维模型的交互方法,可以在检测到用户针对用户界面的目标交互操作的情况下,向为用户界面提供页面数据的服务器发送针对目标交互操作的交互请求。用户界面用于呈现三维模型,三维模型与第二用户终端登录的用户账号预先建立关联关系。之后,接收服务器从第二用户终端获取的流媒体视频。最后,在用户界面上呈现流媒体视频和三维模型。通过在终端设备的同一页面呈现流媒体视频和三维模型,有助于采用流媒体视频向用户呈现三维模型相关的信息,提高了交互方式的多样性。通过多维度信息交互,让用户更加沉静地浏览三维模型,提升用户的浏览时长,有助于满足用户更多元化的交互需求。The interaction method based on the three-dimensional model provided by the above-mentioned embodiments of the present disclosure can send an interaction request for the target interaction operation to a server that provides page data for the user interface when the user's target interaction operation for the user interface is detected. The user interface is used for presenting a three-dimensional model, and the three-dimensional model establishes an association relationship with the user account logged in by the second user terminal in advance. After that, the streaming video obtained by the server from the second user terminal is received. Finally, the streaming video and 3D model are presented on the user interface. By presenting the streaming media video and the 3D model on the same page of the terminal device, it is helpful to use the streaming media video to present information related to the 3D model to the user, and increase the diversity of interaction methods. Through multi-dimensional information interaction, users can browse the three-dimensional model more calmly, improve the user's browsing time, and help meet the users' more diversified interactive needs.
在本实施例的一些可选的实现方式中,第一用户终端还可以执行如下步骤:In some optional implementation manners of this embodiment, the first user terminal may also perform the following steps:
首先,接收服务器发送的模型调整信息,其中,模型调整信息指示使用第二用户终端的用户对呈现于第二用户终端的三维模型的调整操作。调整操作包括以下至少一项:缩放、旋转、移动、视点切换。First, the model adjustment information sent by the server is received, where the model adjustment information indicates an adjustment operation of the user who uses the second user terminal on the three-dimensional model presented on the second user terminal. The adjustment operation includes at least one of the following: zoom, rotate, move, and switch viewpoints.
这里,通常情况下,用户可以对三维模型进行缩放、旋转、移动、视点切换中的至少一项操作。Here, under normal circumstances, the user can perform at least one operation of zooming, rotating, moving, and switching viewpoints on the three-dimensional model.
然后,按照模型调整信息指示的调整操作,对用户界面上呈现的三维模型进行相同调整操作。Then, according to the adjustment operation indicated by the model adjustment information, the same adjustment operation is performed on the three-dimensional model presented on the user interface.
可以理解,上述可选的实现方式中,第二用户终端的用户对三维模型所执行的操作,可以同步到第一用户终端。由此,在第二用户终端采集的流媒体视频与三维模型相关(例如,第二用户中终端的用户对三维模型进行讲解、介绍等)时,方便第一用户终端的用户参考与第二用户终端呈现的相同三维模型,对流媒体视频中的信息进行获取,从而提高了信息获取的针对性。It can be understood that, in the foregoing optional implementation manner, the operations performed by the user of the second user terminal on the three-dimensional model can be synchronized to the first user terminal. Therefore, when the streaming video collected by the second user terminal is related to the three-dimensional model (for example, the user of the second user terminal explains or introduces the three-dimensional model), it is convenient for the user of the first user terminal to refer to the second user The same three-dimensional model presented by the terminal acquires the information in the streaming video, thereby improving the pertinence of information acquisition.
在本实施例的一些可选的实现方式中,第一用户终端还可以执行如下步骤:In some optional implementation manners of this embodiment, the first user terminal may also perform the following steps:
首先,获取用户针对流媒体视频的反馈信息。该反馈信息可以包括但不限于以下至少一项:点赞、评分、评论等等。该反馈信息可以用于表征第一用户终端的用户对第二用户终端的用户的流媒体视频的评价。First, get user feedback on streaming video. The feedback information may include but is not limited to at least one of the following: likes, ratings, comments, and so on. The feedback information may be used to characterize the evaluation of the user of the first user terminal on the streaming video of the user of the second user terminal.
然后,将反馈信息发送至服务器,其中,服务器用于将反馈信息与用户账号建立关联关系。例如,可以采用关联存储的方式,将反馈信息与用户账号建立关联关系。Then, the feedback information is sent to the server, where the server is used to establish an association relationship between the feedback information and the user account. For example, an associative storage method can be used to establish an association relationship between the feedback information and the user account.
可以理解,将反馈信息与用户账号建立关联关系,可以反映第一用户终端的用户对三维模型指示的物体、对第二用户终端的用户的满意程度,进而可以更具针对性地为第一用户终端推送信息。It can be understood that establishing an association relationship between the feedback information and the user account can reflect the user's satisfaction with the object indicated by the three-dimensional model and the user of the second user terminal by the user of the first user terminal, and thus can be more targeted for the first user The terminal pushes information.
进一步参考图9,图9是本公开的第一个基于三维模型的交互方法的又一个实施例的流程900。该基于三维模型的交互方法应用于第一用户终端,第一用户终端呈现有用户界面,该方 法包括:With further reference to FIG. 9, FIG. 9 is a flow 900 of another embodiment of the first three-dimensional model-based interaction method of the present disclosure. The three-dimensional model-based interaction method is applied to a first user terminal, and the first user terminal is presented with a user interface, and the method includes:
步骤910,响应于检测到用户针对用户界面的目标交互操作,向为用户界面提供页面数据的服务器发送针对目标交互操作的交互请求。Step 910: In response to detecting the user's target interaction operation on the user interface, send an interaction request for the target interaction operation to a server that provides page data for the user interface.
步骤920,接收服务器从第二用户终端获取的流媒体视频。Step 920: Receive the streaming video obtained by the server from the second user terminal.
步骤930,在用户界面上呈现流媒体视频和三维模型。Step 930: Present the streaming video and the three-dimensional model on the user interface.
在本实施例中,步骤910至步骤930分别与图7对应实施例中的步骤710至步骤730基本一致,这里不再赘述。In this embodiment, step 910 to step 930 are basically the same as step 710 to step 730 in the embodiment corresponding to FIG. 7, and will not be repeated here.
步骤940,响应于第一用户终端的当前网速值小于或等于预设网速阈值,基于流媒体视频中的各帧语音,对目标用户图像进行调整,生成不同于流媒体视频的新视频。Step 940: In response to the current network speed value of the first user terminal being less than or equal to the preset network speed threshold, adjust the target user image based on each frame of voice in the streaming video to generate a new video different from the streaming video.
在本实施例中,在第一用户终端的当前网速值小于或等于预设网速阈值的情况下,第一用户终端可以基于流媒体视频中的各帧语音,对目标用户图像进行调整,生成新视频。新视频表征目标用户图像指示的用户执行各帧语音指示的动作。目标用户图像指示的用户可以是使用第二用户终端的用户。可选地,新视频可以是基于网络分段发送、即时传输的流媒体视频,也可以是无需基于网络在本地生成的或视频。In this embodiment, when the current network speed value of the first user terminal is less than or equal to the preset network speed threshold, the first user terminal may adjust the target user image based on each frame of voice in the streaming video, Generate a new video. The new video characterizes the actions of the user indicated by the target user's image to perform each frame of voice instructions. The user indicated by the target user image may be a user using the second user terminal. Optionally, the new video may be a streaming video that is sent in segments and instantly transmitted based on the network, or it may be a video that is generated locally without being based on the network.
具体地,第一用户终端可以采用如下方式生成新视频:对于流媒体视频中的每帧语音,将该帧语音输入至预先确定的图像帧生成模型,得到与该帧语音相匹配的、目标用户图像指示的用户的图像。从而将所得到的与流媒体视频中的各帧语音相匹配的各帧图像,以及该各帧语音进行融合,从而得到新视频。与语音相匹配的、目标用户图像指示的用户的图像中用户的动作与该语音相吻合。例如,如果语音为“啊”的音频,并且该音频表征用户处于惊吓状态,那么,与该音频相匹配的目标用户图像指示的用户的图像中的用户的口型可以是发出语音“啊”的口型,动作可以是处于惊吓状态下的动作。Specifically, the first user terminal may generate a new video in the following manner: For each frame of voice in the streaming video, input the frame of voice into a predetermined image frame generation model to obtain a target user that matches the frame of voice The image of the user indicated by the image. Thereby, the obtained frames of images that match each frame of voice in the streaming video and the frames of voice are merged to obtain a new video. The user's action in the user's image indicated by the target user image that matches the voice matches the voice. For example, if the voice is "ah" audio, and the audio indicates that the user is in a startled state, then the mouth shape of the user in the user's image indicated by the target user image matching the audio may be the voice "ah" Lip shape, the action can be an action in a state of fright.
在这里,上述图像帧生成模型可以是采用机器学习算法,基于包括语音帧、目标用户图像和与语音帧相匹配的图像帧的训练样本,训练得到的循环神经网络模型或卷积神经网络模型。针对每个用户可以训练得到一个图像帧生成模型,用以训练该用户的图像帧生成模型的各个训练样本中的目标用户图像可以相同,针对该用户的每个语音帧确定出与该语音帧相匹配的图像帧,进而得到用以训练该用户的图像帧生成模型的训练样本集合。Here, the aforementioned image frame generation model may be a recurrent neural network model or a convolutional neural network model obtained by training using a machine learning algorithm based on training samples including voice frames, target user images, and image frames matching the voice frames. An image frame generation model can be trained for each user, and the target user image in each training sample used to train the user’s image frame generation model can be the same. For each voice frame of the user, it is determined that the voice frame corresponds to the voice frame. Matched image frames, and then obtain a training sample set used to train the image frame generation model of the user.
可选地,图像帧生成模型还可以是关联存储有语音帧、目标用户图像和与语音帧相匹配的图像帧的二维表或数据库。在图像帧生成模型还是关联存储有语音帧、目标用户图像和与语音帧相匹配的图像帧的数据库的情况下,该数据库的每条记录可以包括语音帧、目标用户图像和与语音帧相匹配的图像帧。各条记录中的目标用户图像可以相同,针对该用户的每个语音帧确 定出与该语音帧相匹配的图像帧,进而得到关联存储有语音帧、目标用户图像和与语音帧相匹配的图像帧的数据库,即图像帧生成模型。Optionally, the image frame generation model may also be a two-dimensional table or database that stores the voice frame, the target user image, and the image frame matching the voice frame in association with each other. In the case that the image frame generation model is still associated with a database storing voice frames, target user images and image frames matching the voice frames, each record of the database may include voice frames, target user images and matching voice frames Image frame. The target user image in each record can be the same. For each voice frame of the user, an image frame that matches the voice frame is determined, and then the voice frame, the target user image, and the image that matches the voice frame are associated and stored. Frame database, that is, image frame generation model.
在本实施例的一些可选的实现方式中,第一用户终端还可以通过以下任一方式,确定出目标用户图像:In some optional implementation manners of this embodiment, the first user terminal may also determine the target user image by any of the following methods:
(1)基于流媒体视频中的图像,生成目标用户图像。(1) Based on the image in the streaming video, the target user image is generated.
这里,可以从流媒体视频中的各帧图像中,随机选取一张图像作为目标用户图像,也可以从流媒体视频中的各帧图像中,选取一张面部图像区域与整张图像帧的面积之比大于预设阈值的图像,作为目标用户图像。Here, you can randomly select an image from each frame of the streaming video as the target user image, or you can select a face image area and the area of the entire image frame from each frame of the streaming video The image whose ratio is greater than the preset threshold is regarded as the target user image.
(2)将与用户账号相关联的用户图像确定为目标用户图像。(2) Determine the user image associated with the user account as the target user image.
这里,用户可以通过其所使用的用户账号上传一张图像,作为目标用户图像;也可以在登录其所使用的账号之后,从预先确定的图像集合中选取一张图像,作为目标用户图像。Here, the user can upload an image through the user account he uses as the target user image; or after logging in the account he uses, select an image from a predetermined image set as the target user image.
可以理解,上述可选的实现方式可以实现从流媒体视频中的图像中,自动生成目标用户图像,或者,用户手动设置目标用户图像,从而基于多种目标用户图像的确定方式,使得新视频的生成方式更为多样化。It can be understood that the above-mentioned optional implementation methods can automatically generate a target user image from the images in the streaming video, or the user manually sets the target user image, so that based on multiple target user image determination methods, the new video The generation method is more diversified.
步骤950,采用新视频替代流媒体视频进行呈现。In step 950, the new video is used to replace the streaming video for presentation.
在本实施例中,第一用户终端可以采用新视频替代流媒体视频进行呈现。换言之,在第一用户终端呈现新视频时,可以对流媒体视频进行隐藏(即不再呈现)。In this embodiment, the first user terminal may use a new video to replace the streaming video for presentation. In other words, when the first user terminal presents a new video, the streaming video can be hidden (that is, no longer presented).
需要说明的是,除上面所记载的内容外,本申请实施例还可以包括与图7对应的实施例相同或类似的特征、效果,在此不再赘述。It should be noted that, in addition to the content described above, the embodiment of the present application may also include the same or similar features and effects as the embodiment corresponding to FIG. 7, which will not be repeated here.
从图9中可以看出,本实施例中的基于三维模型的交互方法的流程900在第一用户终端的当前网速值较小(小于或等于预设网速阈值)的情况下,第一用户终端可以在本地生成用以替代流媒体视频呈现的新视频。由此,第一用户终端只需从服务器持续获取语音,而无需持续获取视频,从而降低了对网络资源的占用。在第一用户终端的当前网速值较小的情况下,可以提高第一用户终端的视频呈现的实时性。It can be seen from FIG. 9 that in the process 900 of the interaction method based on the three-dimensional model in this embodiment, when the current network speed value of the first user terminal is small (less than or equal to the preset network speed threshold), the first The user terminal can locally generate a new video to replace the streaming video presentation. Therefore, the first user terminal only needs to continuously obtain voice from the server, but does not need to continuously obtain video, thereby reducing the occupation of network resources. In a case where the current network speed value of the first user terminal is relatively small, the real-time performance of the video presentation of the first user terminal can be improved.
在本实施例的一些可选的实现方式中,在用户界面呈现有新视频(未呈现第二用户终端获取的流媒体视频)的情况下,第一用户终端还可以向上述服务器发送摄像头关闭确认信息。其中,摄像头关闭确认信息用于确定第二用户终端是否关闭摄像头。In some optional implementations of this embodiment, in the case where a new video is presented on the user interface (the streaming video obtained by the second user terminal is not presented), the first user terminal may also send a camera shutdown confirmation to the server. information. Wherein, the camera closing confirmation information is used to determine whether the second user terminal closes the camera.
可以理解,在服务器接收到摄像头关闭确认信息之后,服务器可以向第二用户终端发送用于确定第二用户终端是否关闭摄像头的信息。由此,第二用户终端的用户可以通过关闭摄像头,来降低第二用户终端对网络资源的占用。It can be understood that after the server receives the camera shutdown confirmation information, the server may send to the second user terminal information for determining whether the second user terminal turns off the camera. Therefore, the user of the second user terminal can reduce the occupation of network resources by the second user terminal by turning off the camera.
请继续参考图10,图10是本公开的第一个基于三维模型的交互方法的又一个实施例的流程图。该基于三维模型的交互方法应用于第一用户终端,第一用户终端呈现有用户界面。该基于三维模型的交互方法的流程1000,包括:Please continue to refer to FIG. 10, which is a flowchart of another embodiment of the first three-dimensional model-based interaction method of the present disclosure. The interaction method based on the three-dimensional model is applied to a first user terminal, and the first user terminal presents a user interface. The process 1000 of the interaction method based on the three-dimensional model includes:
步骤1010,响应于检测到用户针对用户界面的目标交互操作,向为用户界面提供页面数据的服务器发送针对目标交互操作的交互请求。用户界面用于呈现三维模型,三维模型与第二用户终端登录的用户账号预先建立关联关系。Step 1010: In response to detecting the user's target interaction operation on the user interface, send an interaction request for the target interaction operation to a server that provides page data for the user interface. The user interface is used for presenting a three-dimensional model, and the three-dimensional model establishes an association relationship with the user account logged in by the second user terminal in advance.
步骤1020,接收服务器从第二用户终端获取的流媒体视频。Step 1020: Receive the streaming video obtained by the server from the second user terminal.
步骤1030,在用户界面上呈现流媒体视频和三维模型。Step 1030: Present the streaming video and the three-dimensional model on the user interface.
在本实施例中,步骤1010至步骤1030分别与图7对应实施例中的步骤710至步骤730基本一致,这里不再赘述。In this embodiment, step 1010 to step 1030 are basically the same as step 710 to step 730 in the embodiment corresponding to FIG. 7, and will not be repeated here.
需要说明的是,在本实施例中,三维模型包括多个子空间场景的三维子模型,多个子空间场景中的子空间场景与预先确定的关键词集合中的关键词相对应。It should be noted that in this embodiment, the three-dimensional model includes three-dimensional sub-models of multiple sub-space scenes, and the sub-space scenes in the multiple sub-space scenes correspond to keywords in a predetermined keyword set.
步骤1040,对流媒体视频中的语音进行语音识别,得到语音识别结果。Step 1040: Perform voice recognition on the voice in the streaming video to obtain a voice recognition result.
在本实施例中,第一用户终端可以对流媒体视频中的语音进行语音识别,得到语音识别结果。In this embodiment, the first user terminal may perform voice recognition on the voice in the streaming video to obtain the voice recognition result.
这里,语音识别结果可以表征流媒体视频中的语音对应的文字。Here, the voice recognition result can represent the text corresponding to the voice in the streaming video.
步骤1050,响应于确定语音识别结果包含关键词集合中的关键词的确定,在用户界面上呈现与语音识别结果包含的关键词相对应的多个子空间场景中的对应子空间场景的三维子模型。Step 1050: In response to the determination that the voice recognition result includes keywords in the keyword set, present on the user interface a three-dimensional sub-model of the corresponding sub-space scene among the multiple sub-space scenes corresponding to the keywords included in the voice recognition result .
在本实施例中,在确定语音识别结果包含关键词集合中的关键词的情况下,第一用户终端可以在上述用户界面上,呈现与语音识别结果包含的关键词相对应的子空间场景的三维子模型。In this embodiment, in the case where it is determined that the voice recognition result includes keywords in the keyword set, the first user terminal may display the subspace scene corresponding to the keywords contained in the voice recognition result on the aforementioned user interface. Three-dimensional sub-model.
作为示例,如果上述三维模型为房屋室内的三维模型。该房屋包括卧室、客厅、厨房、卫生间,共四个子空间场景。也即,上述三维模型包括卧室的三维子模型、客厅的三维子模型、厨房的三维子模型、卫生间的三维子模型。关键词集合包括卧室、客厅、厨房、卫生间。由此,与子空间场景卧室相对应的关键词可以是“卧室”;与子空间场景厨房相对应的关键词可以是“厨房”;与子空间场景客厅相对应的关键词可以是“客厅”;与子空间场景卫生间相对应的关键词可以是“卫生间”。进一步地,作为示例,如果语音识别结果包含关键词“卧室”,那么,第一用户终端可以在上述用户界面上,呈现卧室的三维子模型。As an example, if the above-mentioned three-dimensional model is a three-dimensional model inside a house. The house includes a bedroom, a living room, a kitchen, and a bathroom, with a total of four sub-space scenes. That is, the above-mentioned three-dimensional model includes a three-dimensional sub-model of a bedroom, a three-dimensional sub-model of a living room, a three-dimensional sub-model of a kitchen, and a three-dimensional sub-model of a bathroom. The keyword set includes bedroom, living room, kitchen, bathroom. Therefore, the keyword corresponding to the subspace scene bedroom can be "bedroom"; the keyword corresponding to the subspace scene kitchen can be "kitchen"; the keyword corresponding to the subspace scene living room can be "living room" ; The keyword corresponding to the bathroom in the subspace scene can be "toilet". Further, as an example, if the voice recognition result includes the keyword "bedroom", then the first user terminal may present a three-dimensional sub-model of the bedroom on the aforementioned user interface.
在这里,可以通过切换三维模型的视点,实现呈现与语音识别结果包含的关键词相对应的 子空间场景的三维子模型。Here, by switching the viewpoint of the three-dimensional model, a three-dimensional sub-model that presents the sub-space scene corresponding to the keywords contained in the voice recognition result can be realized.
需要说明的是,除上面所记载的内容外,本申请实施例还可以包括与图7和/或图9对应的实施例相同或类似的特征、效果,在此不再赘述。It should be noted that, in addition to the content described above, the embodiment of the present application may also include the same or similar features and effects as the embodiment corresponding to FIG. 7 and/or FIG. 9, and details are not described herein again.
从图10中可以看出,本实施例中的基于三维模型的交互方法的流程1000中,可以通过语音实现三维模型的视点切换,从而呈现与语音识别结果包含的关键词相对应的子空间场景的三维子模型。由此,提高了三维模型浏览的便利性,提高了所呈现的三维模型与第二用户终端获取的语音之间的匹配性。It can be seen from FIG. 10 that in the process 1000 of the interaction method based on the three-dimensional model in this embodiment, the viewpoint switching of the three-dimensional model can be realized by voice, thereby presenting the subspace scene corresponding to the keywords contained in the voice recognition result The three-dimensional sub-model. As a result, the convenience of browsing the three-dimensional model is improved, and the matching between the presented three-dimensional model and the voice acquired by the second user terminal is improved.
请继续参考图11,示出了根据本公开的第二个基于三维模型的交互方法的一个实施例的流程1100。该基于三维模型的交互方法应用于第二用户终端,第二用户终端登录的用户账号与三维模型预先建立关联关系。该基于三维模型的交互方法包括:Please continue to refer to FIG. 11, which shows a process 1100 of an embodiment of the second three-dimensional model-based interaction method according to the present disclosure. The three-dimensional model-based interaction method is applied to a second user terminal, and the user account logged in by the second user terminal establishes an association relationship with the three-dimensional model in advance. The interactive method based on the 3D model includes:
步骤1110,响应于接收到服务器发送的交互请求,获取流媒体视频。Step 1110: In response to receiving the interactive request sent by the server, obtain the streaming video.
在本实施例中,用户可以使用第二用户终端通过网络与服务器、第一用户终端进行交互。第二用户终端可以是各种电子设备,包括但不限于智能手机、平板电脑、膝上型便携计算机和台式计算机等等。第二用户终端可以安装有各种客户端应用,例如房产交易软件等。In this embodiment, the user can use the second user terminal to interact with the server and the first user terminal through the network. The second user terminal may be various electronic devices, including but not limited to smart phones, tablet computers, laptop computers, desktop computers, and so on. The second user terminal may be installed with various client applications, such as real estate transaction software.
在本实施例中,在接收到服务器发送的交互请求的情况下,获取流媒体视频。In this embodiment, upon receiving the interaction request sent by the server, the streaming video is acquired.
交互请求指示第一用户终端检测到用户针对第一用户终端呈现的用户界面的目标交互操作。示例性地,上述交互请求可以用于指示第一用户终端的用户请求与第二用户终端进行视频通信。用户界面用于呈现三维模型。流媒体视频可以包含图像和/或语音。实践中,第二用户终端的图像获取装置和/或语音获取装置,可以用于获取上述流媒体视频。The interaction request indicates that the first user terminal detects the user's target interaction operation on the user interface presented by the first user terminal. Exemplarily, the aforementioned interaction request may be used to instruct the user of the first user terminal to request video communication with the second user terminal. The user interface is used to present the three-dimensional model. Streaming videos can contain images and/or voice. In practice, the image acquisition device and/or the voice acquisition device of the second user terminal can be used to acquire the aforementioned streaming video.
实践中,在在检测到第一用户终端的用户针对用户界面的目标交互操作的情况下,第一用户终端可以向为用户界面提供页面数据的服务器发送针对目标交互操作的交互请求。用户界面用于呈现三维模型。三维模型与第二用户终端登录的用户账号预先建立关联关系。上述目标交互操作可以是各种用于指示第一用户终端请求与第二用户终端进行交互(信息交互)的操作。作为示例,该目标交互操作可以指示与第二用户终端进行视频通信。In practice, in the case of detecting a target interaction operation of the user of the first user terminal on the user interface, the first user terminal may send an interaction request for the target interaction operation to a server that provides page data for the user interface. The user interface is used to present the three-dimensional model. The three-dimensional model establishes an association relationship with the user account logged in by the second user terminal in advance. The aforementioned target interaction operation may be various operations for instructing the first user terminal to request interaction (information interaction) with the second user terminal. As an example, the target interaction operation may indicate video communication with the second user terminal.
在这里,在执行步骤1110时,第一用户终端的用户界面可以呈现有上述三维模型,也可以未呈现三维模型。Here, when step 1110 is performed, the user interface of the first user terminal may present the above-mentioned three-dimensional model, or may not present the three-dimensional model.
步骤1120,向服务器发送流媒体视频。Step 1120: Send the streaming video to the server.
在本实施例中,第二用户终端可以向服务器发送流媒体视频。服务器用于将流媒体视频发送至第一用户终端,以使第一用户终端在用户界面上呈现流媒体视频和三维模型。In this embodiment, the second user terminal may send the streaming video to the server. The server is used to send the streaming video to the first user terminal, so that the first user terminal presents the streaming video and the three-dimensional model on the user interface.
实践中,服务器可以采用流媒体技术,将第二用户终端采集的图像和/或语音(即流媒体将 视频),持续发送至第一用户终端。流媒体技术是指采用流式传输技术在网络上连续实时播放的媒体格式。这里,第二用户终端可以将其所采集的连续的影像和声音信息经过压缩处理后发送至服务器。由服务器向第一用户终端顺序或实时地传送各个压缩包,让使用第一用户终端的用户一边下载一边观看、收听。In practice, the server can use streaming media technology to continuously send the images and/or voice (that is, streaming video) collected by the second user terminal to the first user terminal. Streaming media technology refers to a media format that uses streaming technology to continuously play on the network in real time. Here, the second user terminal may send the continuous image and sound information collected by it to the server after compression processing. The server transmits each compressed package to the first user terminal sequentially or in real time, so that users who use the first user terminal can watch and listen while downloading.
可选地,服务器可以将第二用户终端采集的流媒体视频发送至第一用户终端,也可以对第二用户终端采集的流媒体视频进行图像处理(例如美颜)、语音处理(例如去噪)、转码、录制、鉴黄等操作后,将处理后的流媒体视频发送至第一用户终端。Optionally, the server may send the streaming video collected by the second user terminal to the first user terminal, and may also perform image processing (such as beauty) and voice processing (such as denoising) on the streaming video collected by the second user terminal. ), after operations such as transcoding, recording, and pornography, send the processed streaming video to the first user terminal.
本公开的上述实施例提供的第二种基于三维模型的交互方法,应用于第二用户终端,第二用户终端登录的用户账号与三维模型预先建立关联关系。第二用户终端可以在接收到服务器发送的交互请求的情况下,确定是否检测到用户针对交互请求的确认操作。交互请求指示第一用户终端检测到用户针对第一用户终端呈现的用户界面的目标交互操作,用户界面用于呈现三维模型。之后,在检测到确认操作的情况下,获取流媒体视频。最后向服务器发送流媒体视频,其中,服务器用于将流媒体视频发送至第一用户终端,以使第一用户终端在用户界面上呈现流媒体视频和三维模型。本公开实施例通过在终端设备的同一页面呈现流媒体视频和三维模型,有助于采用流媒体视频向用户呈现三维模型相关的信息,提高了交互方式的多样性。通过多维度信息交互,让用户更加沉静地浏览三维模型,提升用户的浏览时长,有助于满足用户更多元化的交互需求。The second three-dimensional model-based interaction method provided by the foregoing embodiment of the present disclosure is applied to a second user terminal, and the user account logged in by the second user terminal establishes an association relationship with the three-dimensional model in advance. The second user terminal may determine whether the user's confirmation operation for the interaction request is detected in the case of receiving the interaction request sent by the server. The interaction request indicates that the first user terminal detects the user's target interaction operation on the user interface presented by the first user terminal, and the user interface is used to present a three-dimensional model. Afterwards, if the confirmation operation is detected, the streaming video is obtained. Finally, the streaming video is sent to the server, where the server is used to send the streaming video to the first user terminal, so that the first user terminal presents the streaming video and the three-dimensional model on the user interface. In the embodiments of the present disclosure, by presenting the streaming media video and the three-dimensional model on the same page of the terminal device, it is helpful to use the streaming media video to present information related to the three-dimensional model to the user, thereby increasing the diversity of interaction modes. Through multi-dimensional information interaction, users can browse the three-dimensional model more calmly, improve the user's browsing time, and help meet the users' more diversified interactive needs.
在本实施例的一些可选的实现方式中,上述步骤1110可以包括以下步骤:In some optional implementation manners of this embodiment, the foregoing step 1110 may include the following steps:
首先,在接收到服务器发送的交互请求的情况下,确定是否检测到用户针对交互请求的确认操作响应于检测到确认操作。确认操作表征第二用户终端的用户确认(同意)与第一用户终端进行交互(例如视频通信)。First, in the case of receiving the interaction request sent by the server, it is determined whether the user's confirmation operation for the interaction request is detected in response to the detection of the confirmation operation. The confirmation operation indicates that the user of the second user terminal confirms (agrees) to interact with the first user terminal (for example, video communication).
然后,在检测到确认操作的情况下,获取流媒体视频。Then, when the confirmation operation is detected, the streaming video is obtained.
可以理解,上述可选的实现方式中,在第二用户终端针对交互请求发送的交互确认信息的情况下,第一用户终端可以呈现流媒体视频;而在第二用户终端未发送上述交互确认信息的情况下,第一用户终端则不呈现流媒体视频。由此,可以在获得第二用户终端的用户的允许(例如接通第一用户终端发起的视频通话)后,才在第一用户终端的用户界面上呈现流媒体视频和三维模型。这有助于提高对第二用户终端的用户的隐私保护性,为第二用户终端的用户向第一用户终端的用户呈现流媒体视频提供准备时间。It can be understood that in the foregoing optional implementation manner, in the case of the interaction confirmation information sent by the second user terminal in response to the interaction request, the first user terminal may present the streaming video; and the second user terminal does not send the interaction confirmation information. In the case of, the first user terminal does not present the streaming video. Therefore, the streaming video and the three-dimensional model can be presented on the user interface of the first user terminal only after the permission of the user of the second user terminal is obtained (for example, the video call initiated by the first user terminal is connected). This helps to improve the privacy protection of the user of the second user terminal, and provides preparation time for the user of the second user terminal to present the streaming media video to the user of the first user terminal.
在本实施例的一些可选的实现方式中,在服务器接收到交互请求之后,第二用户终端也可以直接获取流媒体视频,并将流媒体视频通过服务器发送至第一用户终端,而无需第二用户终 端的用户针对交互请求发送的交互确认信息。In some optional implementations of this embodiment, after the server receives the interaction request, the second user terminal may also directly obtain the streaming video, and send the streaming video to the first user terminal through the server, without the need for a second user terminal. 2. Interaction confirmation information sent by the user of the user terminal in response to the interaction request.
可以理解,上述可选的实现方式中,第二用户终端的用户可以处于向其他用户终端的用户拍摄流媒体视频(例如直播)的状态。由此,在服务器接收到交互请求之后,第一用户终端可以随时接收服务器从第二用户终端获取的流媒体视频,从而提高了流媒体视频呈现的实时性。It can be understood that, in the foregoing optional implementation manner, the user of the second user terminal may be in a state of shooting a streaming video (for example, a live broadcast) to users of other user terminals. Thus, after the server receives the interaction request, the first user terminal can receive the streaming video obtained by the server from the second user terminal at any time, thereby improving the real-time performance of the streaming video presentation.
在本实施例的一些可选的实现方式中,在第一用户终端的当前网速值小于或等于预设网速阈值的情况下,第二用户终端可以从服务器接收摄像头关闭确认信息,以及呈现摄像头关闭确认信息。摄像头关闭确认信息用于确定第二用户终端是否关闭摄像头。In some optional implementations of this embodiment, in the case that the current network speed value of the first user terminal is less than or equal to the preset network speed threshold, the second user terminal may receive the camera shutdown confirmation message from the server, and display The camera is turned off confirmation message. The camera close confirmation information is used to determine whether the second user terminal closes the camera.
可以理解,在服务器接收到第一用户终端的当前网速值小于或等于预设网速阈值的信息之后,服务器可以向第二用户终端发送用于确定第二用户终端是否关闭摄像头的信息。由此,第二用户终端的用户可以通过关闭摄像头,来降低第二用户终端对网络资源的占用。It can be understood that after the server receives the information that the current network speed value of the first user terminal is less than or equal to the preset network speed threshold, the server may send to the second user terminal information for determining whether the second user terminal turns off the camera. Therefore, the user of the second user terminal can reduce the occupation of network resources by the second user terminal by turning off the camera.
在本实施例中的一些可选的实现方式中,在检测到用户对呈现于第二用户终端的三维模型的调整操作的情况下,第二用户终端可以向服务器发送指示调整操作的模型调整信息,以使服务器控制第一用户终端按照模型调整信息指示的调整操作,对用户界面上呈现的三维模型进行相同调整操作。调整操作包括以下至少一项:缩放、旋转、移动、视点切换。In some optional implementation manners in this embodiment, in a case where the user's adjustment operation on the three-dimensional model presented on the second user terminal is detected, the second user terminal may send model adjustment information indicating the adjustment operation to the server , So that the server controls the first user terminal to perform the same adjustment operation on the three-dimensional model presented on the user interface according to the adjustment operation indicated by the model adjustment information. The adjustment operation includes at least one of the following: zoom, rotate, move, and switch viewpoints.
这里,通常情况下,用户可以对三维模型进行缩放、旋转、移动、视点切换中的至少一项操作。Here, under normal circumstances, the user can perform at least one operation of zooming, rotating, moving, and switching viewpoints on the three-dimensional model.
可以理解,上述可选的实现方式中,第二用户终端的用户对三维模型所执行的操作,可以同步到第一用户终端。由此,在第二用户终端采集的流媒体视频与三维模型相关(例如,第二用户中终端的用户对三维模型进行讲解、介绍等)时,方便第一用户终端的用户参考与第二用户终端呈现的相同三维模型,对流媒体视频中的信息进行获取,从而提高了信息获取的针对性。It can be understood that, in the foregoing optional implementation manner, the operations performed by the user of the second user terminal on the three-dimensional model can be synchronized to the first user terminal. Therefore, when the streaming video collected by the second user terminal is related to the three-dimensional model (for example, the user of the second user terminal explains or introduces the three-dimensional model), it is convenient for the user of the first user terminal to refer to the second user The same three-dimensional model presented by the terminal acquires the information in the streaming video, thereby improving the pertinence of information acquisition.
在本实施例中的一些可选的实现方式中,在从服务器接收到用户对呈现于第一用户终端的三维模型的调整操作的情况下,第二用户终端可以按照模型调整信息指示的调整操作,对第二用户终端呈现的三维模型进行相同调整操作。调整操作包括以下至少一项:缩放、旋转、移动、视点切换。In some optional implementation manners in this embodiment, in the case that the user's adjustment operation on the three-dimensional model presented on the first user terminal is received from the server, the second user terminal may follow the adjustment operation indicated by the model adjustment information , Perform the same adjustment operation on the three-dimensional model presented by the second user terminal. The adjustment operation includes at least one of the following: zoom, rotate, move, and switch viewpoints.
这里,通常情况下,用户可以对三维模型进行缩放、旋转、移动、视点切换中的至少一项操作。Here, under normal circumstances, the user can perform at least one operation of zooming, rotating, moving, and switching viewpoints on the three-dimensional model.
可以理解,上述可选的实现方式中,第一用户终端的用户对三维模型所执行的操作,可以同步到第二用户终端。由此,可以方便第一用户终端的用户参考与第二用户终端呈现的相同三维模型,对流媒体视频中的信息进行获取,从而提高了信息获取的针对性。It can be understood that, in the foregoing optional implementation manner, the operations performed by the user of the first user terminal on the three-dimensional model can be synchronized to the second user terminal. In this way, it is convenient for the user of the first user terminal to refer to the same three-dimensional model presented by the second user terminal to obtain information in the streaming video, thereby improving the pertinence of information acquisition.
在本实施例中的一些可选的实现方式中,在接收到服务器发送的、使用第一用户终端的用 户针对流媒体视频的反馈信息的情况下,上述第二用户终端可以执行与反馈信息相匹配的操作。该反馈信息可以包括但不限于以下至少一项:点赞、评分、评论等等。该反馈信息可以用于表征第一用户终端的用户对第二用户终端的用户的流媒体视频的评价。In some optional implementation manners in this embodiment, in the case of receiving feedback information for the streaming video from the user using the first user terminal sent by the server, the second user terminal may perform the same method as the feedback information. Matching operation. The feedback information may include but is not limited to at least one of the following: likes, ratings, comments, and so on. The feedback information may be used to characterize the evaluation of the user of the first user terminal on the streaming video of the user of the second user terminal.
作为示例,如果第一用户终端的用户针对流媒体视频的反馈信息为点赞,那么,第二用户终端可以呈现与该反馈信息相匹配的操作,例如呈现“XX为您点了一赞!”。As an example, if the feedback information of the user of the first user terminal for the streaming video is likes, then the second user terminal may present an operation that matches the feedback information, for example, “XX gave you a like!” .
可以理解,上述可选的实现方式可以提高互动的真实性、多样性。It can be understood that the foregoing optional implementation manners can improve the authenticity and diversity of interaction.
进一步参考图12,图12是本公开的第二个基于三维模型的交互方法的又一个实施例的流程1200,该基于三维模型的交互方法应用于第一用户终端,第一用户终端呈现有用户界面,该方法包括:With further reference to FIG. 12, FIG. 12 is a flow 1200 of another embodiment of the second three-dimensional model-based interaction method of the present disclosure. The three-dimensional model-based interaction method is applied to a first user terminal, and the first user terminal presents a user Interface, the method includes:
步骤1210,响应于接收到服务器发送的交互请求,获取流媒体视频。Step 1210: In response to receiving the interactive request sent by the server, obtain the streaming video.
步骤1220,向服务器发送流媒体视频。Step 1220: Send the streaming video to the server.
在本实施例中,步骤1210至步骤1220分别与图11对应实施例中的步骤1110至步骤1120基本一致,这里不再赘述。In this embodiment, step 1210 to step 1220 are basically the same as step 1110 to step 1120 in the embodiment corresponding to FIG. 11, and will not be repeated here.
需要说明的是,在本实施例中,三维模型包括多个子空间场景的三维子模型,多个子空间场景中的子空间场景与预先确定的关键词集合中的关键词相对应。It should be noted that in this embodiment, the three-dimensional model includes three-dimensional sub-models of multiple sub-space scenes, and the sub-space scenes in the multiple sub-space scenes correspond to keywords in a predetermined keyword set.
步骤1230,对第一用户终端获取的语音进行语音识别,得到语音识别结果。Step 1230: Perform voice recognition on the voice acquired by the first user terminal to obtain a voice recognition result.
在本实施例中,第二用户终端可以对第一用户终端获取的语音进行语音识别,得到语音识别结果。In this embodiment, the second user terminal may perform voice recognition on the voice acquired by the first user terminal to obtain a voice recognition result.
这里,语音识别结果可以表征流媒体视频中的语音对应的文字。Here, the voice recognition result can represent the text corresponding to the voice in the streaming video.
步骤1240,响应于确定语音识别结果包含关键词集合中的关键词的确定,在用户界面上呈现与语音识别结果包含的关键词相对应的多个子空间场景中的对应子空间场景的三维子模型。 Step 1240, in response to the determination that the voice recognition result contains keywords in the keyword set, present on the user interface a three-dimensional sub-model of the corresponding sub-space scene among the multiple sub-space scenes corresponding to the keywords contained in the voice recognition result .
在本实施例中,在确定语音识别结果包含关键词集合中的关键词的情况下,第二用户终端可以在用户界面上呈现与语音识别结果包含的关键词相对应的子空间场景的三维子模型。In this embodiment, in the case where it is determined that the voice recognition result includes keywords in the keyword set, the second user terminal may present on the user interface a three-dimensional subspace scene corresponding to the keywords contained in the voice recognition result. Model.
作为示例,如果上述三维模型为房屋室内的三维模型。该房屋包括卧室、客厅、厨房、卫生间,共四个子空间场景,也即上述三维模型包括卧室的三维子模型、客厅的三维子模型、厨房的三维子模型、卫生间的三维子模型。关键词集合包括卧室、客厅、厨房、卫生间。由此,与子空间场景卧室相对应的关键词可以是“卧室”;与子空间场景厨房相对应的关键词可以是“厨房”;与子空间场景客厅相对应的关键词可以是“客厅”;与子空间场景卫生间相对应的关键词可以是“卫生间”。进一步地,作为示例,如果语音识别结果包含关键词“卧室”,那么,第二 用户终端可以在上述用户界面上,呈现卧室的三维子模型。As an example, if the above-mentioned three-dimensional model is a three-dimensional model inside a house. The house includes a bedroom, a living room, a kitchen, and a bathroom, with a total of four sub-space scenes, that is, the above-mentioned three-dimensional model includes a three-dimensional sub-model of the bedroom, a three-dimensional sub-model of the living room, a three-dimensional sub-model of the kitchen, and a three-dimensional sub-model of the bathroom. The keyword set includes bedroom, living room, kitchen, bathroom. Therefore, the keyword corresponding to the subspace scene bedroom can be "bedroom"; the keyword corresponding to the subspace scene kitchen can be "kitchen"; the keyword corresponding to the subspace scene living room can be "living room" ; The keyword corresponding to the bathroom in the subspace scene can be "toilet". Further, as an example, if the voice recognition result includes the keyword "bedroom", then the second user terminal may present a three-dimensional sub-model of the bedroom on the aforementioned user interface.
在这里,可以通过切换三维模型的视点,实现呈现与语音识别结果包含的关键词相对应的子空间场景的三维子模型。Here, by switching the viewpoint of the three-dimensional model, a three-dimensional sub-model that presents the sub-space scene corresponding to the keywords contained in the voice recognition result can be realized.
需要说明的是,除上面所记载的内容外,本申请实施例还可以包括与图11对应的实施例相同或类似的特征、效果,在此不再赘述。It should be noted that, in addition to the content described above, the embodiment of the present application may also include the same or similar features and effects as the embodiment corresponding to FIG. 11, and details are not described herein again.
从图12中可以看出,本实施例中的基于三维模型的交互方法的流程1200中,可以通过语音实现三维模型的视点切换,从而呈现与语音识别结果包含的关键词相对应的子空间场景的三维子模型。由此,提高了三维模型浏览的便利性,提高了所呈现的三维模型与第二用户终端获取的语音之间的匹配性。It can be seen from FIG. 12 that in the process 1200 of the interaction method based on the three-dimensional model in this embodiment, the viewpoint switching of the three-dimensional model can be realized by voice, thereby presenting the subspace scene corresponding to the keywords contained in the voice recognition result. The three-dimensional sub-model. As a result, the convenience of browsing the three-dimensional model is improved, and the matching between the presented three-dimensional model and the voice acquired by the second user terminal is improved.
进一步参考图13,作为对上述第一个基于三维模型的交互方法的实现,本公开提供了一种基于三维模型的交互装置的一个实施例,该装置实施例与图7、9、10所示的方法实施例相对应,除下面所记载的特征外,该装置实施例还可以包括与图7、9、10所示的方法实施例相同或相应的特征,以及产生与图7、9、10所示的方法实施例相同或相应的效果。With further reference to FIG. 13, as an implementation of the above-mentioned first interaction method based on a three-dimensional model, the present disclosure provides an embodiment of an interaction device based on a three-dimensional model. In addition to the features described below, the device embodiment may also include the same or corresponding features as the method embodiment shown in FIGS. 7, 9, and 10, and produce the same as those in the method embodiment shown in FIGS. The method embodiments shown have the same or corresponding effects.
如图13所示,本实施例的基于三维模型的交互装置1300设置于第一用户终端,第一用户终端呈现有用户界面。该装置1300包括:第一发送单元1310,被配置成响应于检测到用户针对用户界面的目标交互操作,向为用户界面提供页面数据的服务器发送针对目标交互操作的交互请求,其中,用户界面用于呈现三维模型,三维模型与第二用户终端登录的用户账号预先建立关联关系;第一接收单元1320,被配置成接收服务器从第二用户终端获取的流媒体视频;第一呈现单元1330,被配置成在用户界面上呈现流媒体视频和三维模型。As shown in FIG. 13, the interaction apparatus 1300 based on the three-dimensional model of this embodiment is set in a first user terminal, and the first user terminal presents a user interface. The device 1300 includes: a first sending unit 1310 configured to send an interaction request for the target interaction operation to a server that provides page data for the user interface in response to detecting a user's target interaction operation on the user interface, wherein the user interface For presenting the three-dimensional model, the three-dimensional model is pre-associated with the user account logged in the second user terminal; the first receiving unit 1320 is configured to receive the streaming video obtained by the server from the second user terminal; the first presenting unit 1330 is It is configured to present streaming video and 3D models on the user interface.
在本实施例中,在检测到用户针对用户界面的目标交互操作的情况下,基于三维模型的交互装置1300的第一发送单元1310可以向为用户界面提供页面数据的服务器发送针对目标交互操作的交互请求。用户界面用于呈现三维模型,三维模型与第二用户终端登录的用户账号预先建立关联关系。In this embodiment, in the case of detecting the user's target interaction operation on the user interface, the first sending unit 1310 of the interactive device 1300 based on the three-dimensional model may send the information about the target interaction operation to the server that provides page data for the user interface. Interactive request. The user interface is used for presenting a three-dimensional model, and the three-dimensional model establishes an association relationship with the user account logged in by the second user terminal in advance.
在本实施例中,第一接收单元1320可以接收服务器从第二用户终端获取的流媒体视频。In this embodiment, the first receiving unit 1320 may receive the streaming video obtained by the server from the second user terminal.
在本实施例中,第一呈现单元1330可以在用户界面上呈现流媒体视频和三维模型。In this embodiment, the first presentation unit 1330 may present the streaming video and the three-dimensional model on the user interface.
在本实施例的一些可选的实现方式中,第一接收单元进一步被配置成:响应于服务器接收到第二用户终端针对交互请求发送的交互确认信息,接收服务器从第二用户终端获取的流媒体视频。In some optional implementation manners of this embodiment, the first receiving unit is further configured to: in response to the server receiving the interaction confirmation information sent by the second user terminal in response to the interaction request, receive the stream obtained by the server from the second user terminal. Media video.
在本实施例的一些可选的实现方式中,该装置1300还包括:第一调整单元(图中未示出),被配置成响应于第一用户终端的当前网速值小于或等于预设网速阈值,基于流媒体视频中的各 帧语音,对目标用户图像进行调整,生成新视频,其中,新视频表征目标用户图像指示的用户执行各帧语音指示的动作;第二呈现单元(图中未示出),被配置成采用新视频替代流媒体视频进行呈现。In some optional implementation manners of this embodiment, the device 1300 further includes: a first adjustment unit (not shown in the figure) configured to respond to the current network speed value of the first user terminal being less than or equal to a preset Network speed threshold, based on each frame of voice in the streaming video, adjust the target user's image to generate a new video, where the new video characterizes the user indicated by the target user's image to perform the actions of each frame of voice instruction; the second presentation unit (Figure Not shown in ), is configured to use new video instead of streaming video for presentation.
在本实施例的一些可选的实现方式中,该装置1300还包括:第一生成单元(图中未示出),被配置成基于流媒体视频中的图像,生成目标用户图像;或者,第一确定单元(图中未示出),被配置成将与用户账号相关联的用户图像确定为目标用户图像。In some optional implementation manners of this embodiment, the device 1300 further includes: a first generating unit (not shown in the figure) configured to generate a target user image based on an image in the streaming video; or A determining unit (not shown in the figure) is configured to determine the user image associated with the user account as the target user image.
在本实施例的一些可选的实现方式中,该装置1300还包括:第二发送单元(图中未示出),被配置成响应于用户界面呈现有新视频,向服务器发送摄像头关闭确认信息,其中,摄像头关闭确认信息用于确定第二用户终端是否关闭摄像头。In some optional implementation manners of this embodiment, the device 1300 further includes: a second sending unit (not shown in the figure), configured to send camera shutdown confirmation information to the server in response to a new video presented on the user interface , Wherein the camera closing confirmation information is used to determine whether the second user terminal closes the camera.
在本实施例的一些可选的实现方式中,第一接收单元进一步被配置成:将第一用户终端的当前网速值发送至服务器;接收服务器从第二用户终端获取并发送的流媒体视频,该流媒体视频具有与当前网速值相匹配的分辨率。In some optional implementation manners of this embodiment, the first receiving unit is further configured to: send the current network speed value of the first user terminal to the server; receive the streaming video that the server obtains and sends from the second user terminal , The streaming video has a resolution that matches the current network speed value.
在本实施例的一些可选的实现方式中,该装置1300还包括:第二接收单元(图中未示出),被配置成接收服务器发送的模型调整信息,其中,模型调整信息指示使用第二用户终端的用户对呈现于第二用户终端的三维模型的调整操作,调整操作包括以下至少一项:缩放、旋转、移动、视点切换;第二调整单元(图中未示出),被配置成按照模型调整信息指示的调整操作,对用户界面上呈现的三维模型进行相同调整操作。In some optional implementation manners of this embodiment, the device 1300 further includes: a second receiving unit (not shown in the figure) configured to receive model adjustment information sent by the server, wherein the model adjustment information indicates the use of the first 2. The user of the user terminal adjusts the three-dimensional model presented on the second user terminal. The adjustment operation includes at least one of the following: zoom, rotate, move, and switch viewpoints; the second adjustment unit (not shown in the figure) is configured According to the adjustment operation indicated by the model adjustment information, the same adjustment operation is performed on the three-dimensional model presented on the user interface.
在本实施例的一些可选的实现方式中,三维模型包括多个子空间场景的三维子模型,多个子空间场景中的子空间场景与预先确定的关键词集合中的关键词相对应;以及,该装置1300还包括:第一识别单元(图中未示出),被配置成对流媒体视频中的语音进行语音识别,得到语音识别结果;第三呈现单元(图中未示出),被配置成响应于确定语音识别结果包含关键词集合中的关键词,在用户界面上呈现与语音识别结果包含的关键词相对应的子空间场景的三维子模型。In some optional implementations of this embodiment, the three-dimensional model includes three-dimensional sub-models of multiple sub-space scenes, and the sub-space scenes in the multiple sub-space scenes correspond to keywords in a predetermined keyword set; and, The device 1300 also includes: a first recognition unit (not shown in the figure), configured to perform voice recognition on the voice in the streaming video to obtain a voice recognition result; and a third presentation unit (not shown in the figure), configured In response to determining that the voice recognition result contains the keywords in the keyword set, a three-dimensional sub-model of the subspace scene corresponding to the keywords contained in the voice recognition result is presented on the user interface.
在本实施例的一些可选的实现方式中,该装置1300还包括:第一获取单元(图中未示出),被配置成获取用户针对流媒体视频的反馈信息;第三发送单元(图中未示出),被配置成将反馈信息发送至服务器,其中,服务器用于将反馈信息与用户账号建立关联关系。In some optional implementations of this embodiment, the device 1300 further includes: a first acquiring unit (not shown in the figure), configured to acquire user feedback information for streaming media videos; and a third sending unit (not shown in the figure). Not shown in ), is configured to send feedback information to a server, where the server is used to establish an association relationship between the feedback information and the user account.
本公开的上述实施例提供的基于三维模型的交互装置设置于第一用户终端,第一用户终端呈现有用户界面。该装置1300中,在检测到用户针对用户界面的目标交互操作的情况下,第一发送单元1310可以向为用户界面提供页面数据的服务器发送针对目标交互操作的交互请求,其中,用户界面用于呈现三维模型,三维模型与第二用户终端登录的用户账号预先建立关 联关系,然后,第一接收单元1320接收服务器从第二用户终端获取的流媒体视频,最后,第一呈现单元1330在用户界面上呈现流媒体视频和三维模型。由此,可以在终端设备的同一页面呈现流媒体视频和三维模型,有助于采用流媒体视频向用户呈现三维模型相关的信息,提高了交互方式的多样性,可以通过多维度信息交互,让用户更加沉静的浏览三维模型,提升用户的浏览时长,有助于满足用户更多元化的交互需求。The interaction device based on the three-dimensional model provided by the above-mentioned embodiments of the present disclosure is set in a first user terminal, and the first user terminal presents a user interface. In the apparatus 1300, in the case of detecting a user's target interaction operation on the user interface, the first sending unit 1310 may send an interaction request for the target interaction operation to a server that provides page data for the user interface, where the user interface is used for The three-dimensional model is presented. The three-dimensional model is pre-associated with the user account logged in the second user terminal. Then, the first receiving unit 1320 receives the streaming video obtained by the server from the second user terminal. Finally, the first presenting unit 1330 displays on the user interface Streaming videos and 3D models are presented on the website. As a result, streaming media videos and 3D models can be presented on the same page of the terminal device, which helps to use streaming media videos to present information related to the 3D model to users, which improves the diversity of interaction methods. Users browse the 3D model more calmly, increase the user's browsing time, and help meet users' more diversified interactive needs.
进一步参考图14,作为对上述第二个基于三维模型的交互方法的实现,本公开提供了第二种基于三维模型的交互装置的一个实施例,该装置实施例与图11、12所示的方法实施例相对应,除下面所记载的特征外,该装置实施例还可以包括与图11、12所示的方法实施例相同或相应的特征,以及产生与图11、12所示的方法实施例相同或相应的效果。With further reference to FIG. 14, as an implementation of the above-mentioned second interaction method based on a three-dimensional model, the present disclosure provides an embodiment of a second interaction device based on a three-dimensional model. Corresponding to the method embodiment, in addition to the features described below, the device embodiment may also include the same or corresponding features as the method embodiment shown in Figs. 11 and 12, and the method implementation shown in Figs. The same or corresponding effect.
如图14所示,本实施例的基于三维模型的交互装置1400设置于第二用户终端。该装置1400包括:第二确定单元1410,被配置成响应于接收到服务器发送的交互请求,获取流媒体视频,其中,交互请求指示第一用户终端检测到用户针对第一用户终端呈现的用户界面的目标交互操作,用户界面用于呈现三维模型,三维模型与所述第二用户终端登录的用户账号预先建立关联关系;第四发送单元1420,被配置成向服务器发送流媒体视频,其中,服务器用于将流媒体视频发送至第一用户终端,以使第一用户终端在用户界面上呈现流媒体视频和三维模型。As shown in FIG. 14, the interaction device 1400 based on the three-dimensional model of this embodiment is set in the second user terminal. The device 1400 includes: a second determining unit 1410 configured to obtain a streaming video in response to receiving an interaction request sent by a server, wherein the interaction request indicates that the first user terminal detects that the user interface presented to the first user terminal The user interface is used to present a three-dimensional model, and the three-dimensional model has a pre-established association relationship with the user account logged in the second user terminal; the fourth sending unit 1420 is configured to send streaming video to the server, where the server It is used to send the streaming video to the first user terminal, so that the first user terminal presents the streaming video and the three-dimensional model on the user interface.
在本实施例中,在接收到服务器发送的交互请求的情况下,第二确定单元1410可以获取流媒体视频。其中,交互请求指示第一用户终端检测到用户针对第一用户终端呈现的用户界面的目标交互操作,用户界面用于呈现三维模型。In this embodiment, upon receiving the interaction request sent by the server, the second determining unit 1410 may obtain the streaming video. The interaction request indicates that the first user terminal detects a user's target interaction operation on the user interface presented by the first user terminal, and the user interface is used to present a three-dimensional model.
在本实施例中,第四发送单元1420可以成向服务器发送流媒体视频,其中,服务器用于将流媒体视频发送至第一用户终端,以使第一用户终端在用户界面上呈现流媒体视频和三维模型。In this embodiment, the fourth sending unit 1420 may be configured to send the streaming video to the server, where the server is used to send the streaming video to the first user terminal, so that the first user terminal displays the streaming video on the user interface. And three-dimensional models.
在本实施例的一些可选的实现方式中,第二确定单元1410进一步被配置成:响应于接收到服务器发送的交互请求,确定是否检测到用户针对交互请求的确认操作;响应于检测到确认操作,获取流媒体视频。In some optional implementations of this embodiment, the second determining unit 1410 is further configured to: in response to receiving the interaction request sent by the server, determine whether a confirmation operation of the user for the interaction request is detected; in response to detecting the confirmation Operation to obtain streaming video.
在本实施例的一些可选的实现方式中,该装置1400还包括:第三接收单元(图中未示出),被配置成响应于第一用户终端的当前网速值小于或等于预设网速阈值,从服务器接收摄像头关闭确认信息,以及呈现摄像头关闭确认信息,其中,摄像头关闭确认信息用于确定第二用户终端是否关闭摄像头。In some optional implementations of this embodiment, the device 1400 further includes: a third receiving unit (not shown in the figure), configured to respond to the current network speed value of the first user terminal being less than or equal to a preset The network speed threshold, receiving camera closing confirmation information from the server, and presenting camera closing confirmation information, where the camera closing confirmation information is used to determine whether the second user terminal closes the camera.
在本实施例的一些可选的实现方式中,该装置1400还包括:第五发送单元(图中未示出),被配置成响应于从服务器接收到用户对呈现于第二用户终端的三维模型的调整操作,按照模型 调整信息指示的调整操作,对用户界面上呈现的三维模型进行相同调整操作,其中,调整操作包括以下至少一项:缩放、旋转、移动、视点切换。In some optional implementation manners of this embodiment, the device 1400 further includes: a fifth sending unit (not shown in the figure), configured to respond to receiving from the server the three-dimensional information presented on the second user terminal by the user. The adjustment operation of the model is to perform the same adjustment operation on the three-dimensional model presented on the user interface according to the adjustment operation indicated by the model adjustment information, where the adjustment operation includes at least one of the following: zoom, rotate, move, and switch viewpoints.
在本实施例的一些可选的实现方式中,该装置1400还包括:第五发送单元(图中未示出),被配置成响应于检测到用户对呈现于第二用户终端的三维模型的调整操作,向服务器发送指示调整操作的模型调整信息,以使服务器控制第一用户终端按照模型调整信息指示的调整操作,对用户界面上呈现的三维模型进行相同调整操作,其中,调整操作包括以下至少一项:缩放、旋转、移动、视点切换。In some optional implementation manners of this embodiment, the device 1400 further includes: a fifth sending unit (not shown in the figure), configured to respond to detecting that the user has received the three-dimensional model presented on the second user terminal The adjustment operation is to send model adjustment information indicating the adjustment operation to the server, so that the server controls the first user terminal to perform the same adjustment operation on the three-dimensional model presented on the user interface according to the adjustment operation indicated by the model adjustment information. The adjustment operation includes the following At least one item: zoom, rotate, move, viewpoint switch.
在本实施例的一些可选的实现方式中,三维模型包括多个子空间场景的三维子模型,多个子空间场景中的子空间场景与预先确定的关键词集合中的关键词相对应;以及,该装置1400还包括:第二识别单元(图中未示出),被配置成对第一用户终端获取的语音进行语音识别,得到语音识别结果;第四呈现单元(图中未示出),被配置成响应于确定语音识别结果包含关键词集合中的关键词,在用户界面上呈现与语音识别结果包含的关键词相对应的子空间场景的三维子模型。In some optional implementations of this embodiment, the three-dimensional model includes three-dimensional sub-models of multiple sub-space scenes, and the sub-space scenes in the multiple sub-space scenes correspond to keywords in a predetermined keyword set; and, The device 1400 further includes: a second recognition unit (not shown in the figure), configured to perform voice recognition on the voice acquired by the first user terminal to obtain a voice recognition result; and a fourth presentation unit (not shown in the figure), It is configured to, in response to determining that the voice recognition result contains the keywords in the keyword set, present on the user interface a three-dimensional sub-model of the subspace scene corresponding to the keywords contained in the voice recognition result.
在本实施例的一些可选的实现方式中,该装置1400还包括:执行单元(图中未示出),被配置成响应于接收到服务器发送的、使用第一用户终端的用户针对流媒体视频的反馈信息,执行与反馈信息相匹配的操作。In some optional implementation manners of this embodiment, the apparatus 1400 further includes: an execution unit (not shown in the figure), configured to respond to receiving a stream media message sent by the server and sent by the user using the first user terminal. Video feedback information, perform operations that match the feedback information.
本公开的上述实施例提供的基于三维模型的交互装置设置于第二用户终端,第二用户终端登录的用户账号与三维模型预先建立关联关系,该装置1400中,在接收到服务器发送的交互请求的情况下,第二确定单元1410可以获取流媒体视频,其中,交互请求指示第一用户终端检测到用户针对第一用户终端呈现的用户界面的目标交互操作,用户界面用于呈现三维模型,之后,第四发送单元1420可以向服务器发送流媒体视频,其中,服务器用于将流媒体视频发送至第一用户终端,以使第一用户终端在用户界面上呈现流媒体视频和三维模型。由此,可以在终端设备的同一页面呈现流媒体视频和三维模型,有助于采用流媒体视频向用户呈现三维模型相关的信息,提高了交互方式的多样性,可以通过多维度信息交互,让用户更加沉静的浏览三维模型,提升用户的浏览时长,有助于满足用户更多元化的交互需求。The interaction device based on the three-dimensional model provided by the above-mentioned embodiment of the present disclosure is set in the second user terminal, and the user account logged in by the second user terminal establishes an association relationship with the three-dimensional model in advance. In the device 1400, upon receiving the interaction request sent by the server In the case of, the second determining unit 1410 may obtain the streaming video, where the interaction request indicates that the first user terminal detects the user's target interaction operation on the user interface presented by the first user terminal, and the user interface is used to present the three-dimensional model. The fourth sending unit 1420 may send the streaming video to the server, where the server is used to send the streaming video to the first user terminal, so that the first user terminal can present the streaming video and the three-dimensional model on the user interface. As a result, streaming videos and 3D models can be presented on the same page of the terminal device, which helps to use streaming videos to present information related to the 3D model to users, which improves the diversity of interaction methods. Users browse the 3D model more calmly, increase the user's browsing time, and help meet users' more diversified interactive needs.
请继续参考图15,图15为本公开的基于三维模型的交互系统的一个实施例1500的交互示意图。该基于三维模型的交互系统包括第一用户终端、第二用户终端和服务器,第一用户终端呈现有用户界面,服务器与第一用户终端、第二用户终端分别通信连接。Please continue to refer to FIG. 15, which is a schematic diagram of interaction of an embodiment 1500 of the interactive system based on a three-dimensional model of the present disclosure. The interactive system based on the three-dimensional model includes a first user terminal, a second user terminal, and a server. The first user terminal presents a user interface, and the server is respectively communicatively connected with the first user terminal and the second user terminal.
图15所示,该基于三维模型的交互系统中的第一用户终端、第二用户终端和服务器可以执行如下步骤:As shown in FIG. 15, the first user terminal, the second user terminal, and the server in the interactive system based on the three-dimensional model can perform the following steps:
步骤1501,第一用户终端检测到用户针对用户界面的目标交互操作。Step 1501: The first user terminal detects the user's target interaction operation on the user interface.
在本实施例中,第一用户终端检测到用户针对用户界面的目标交互操作。用户界面用于呈现三维模型,三维模型与第二用户终端登录的用户账号预先建立关联关系。In this embodiment, the first user terminal detects the user's target interaction operation on the user interface. The user interface is used for presenting a three-dimensional model, and the three-dimensional model establishes an association relationship with the user account logged in by the second user terminal in advance.
步骤1502,第一用户终端向服务器发送针对目标交互操作的交互请求。Step 1502: The first user terminal sends an interaction request for the target interaction operation to the server.
在本实施例中,第一用户终端可以向服务器发送针对目标交互操作的交互请求。In this embodiment, the first user terminal may send an interaction request for a target interaction operation to the server.
步骤1503,第二用户终端获取流媒体视频。Step 1503: The second user terminal obtains the streaming video.
在本实施例中,第二用户终端可以获取流媒体视频。In this embodiment, the second user terminal can obtain the streaming video.
步骤1504,第二用户终端向服务器发送流媒体视频。Step 1504: The second user terminal sends the streaming video to the server.
在本实施例中,第二用户终端可以向服务器发送流媒体视频。In this embodiment, the second user terminal may send the streaming video to the server.
步骤1505,服务器向第一用户终端发送流媒体视频。Step 1505: The server sends the streaming video to the first user terminal.
在本实施例中,服务器可以向第一用户终端发送流媒体视频。In this embodiment, the server may send the streaming video to the first user terminal.
步骤1506,第一用户终端在用户界面上呈现流媒体视频和三维模型。Step 1506: The first user terminal presents the streaming video and the three-dimensional model on the user interface.
在本实施例中,第一用户终端可以在用户界面上呈现流媒体视频和三维模型。In this embodiment, the first user terminal may present the streaming video and the three-dimensional model on the user interface.
在本实施例中,在不冲突的前提下,除本实施例中上述记载的内容之外,步骤1501至步骤1506中的技术特征还可以参考上述第一种基于三维模型的交互方法的各个实施例、第二种基于三维模型的交互方法的各个实施例,以及第三种基于三维模型的交互方法中的各个实施例中的技术特征进行解释。并且,该实施例还可以包括与上述基于三维模型的交互方法实施例相同或相应的特征,以及产生与其相同或相应的效果,在此不再赘述。In this embodiment, on the premise of no conflict, in addition to the content described in this embodiment, the technical features in step 1501 to step 1506 can also refer to the implementation of the first three-dimensional model-based interaction method described above. The technical features in each embodiment of the second three-dimensional model-based interaction method, and the third three-dimensional model-based interaction method will be explained. Moreover, this embodiment may also include the same or corresponding features as the above-mentioned embodiment of the interaction method based on the three-dimensional model, and produce the same or corresponding effects, which will not be repeated here.
本公开的上述实施例提供的基于三维模型的交互系统包括第一用户终端、第二用户终端和服务器,第一用户终端呈现有用户界面,服务器与第一用户终端、第二用户终端分别通信连接。其中:第一用户终端被配置成:响应于检测到用户针对用户界面的目标交互操作,向服务器发送针对目标交互操作的交互请求,其中,用户界面用于呈现三维模型,三维模型与第二用户终端登录的用户账号预先建立关联关系;第二用户终端被配置成:获取流媒体视频;向服务器发送流媒体视频;服务器还被配置成:向第一用户终端发送流媒体视频;第一用户终端被配置成:在用户界面上呈现流媒体视频和三维模型。由此,可以在终端设备的同一页面呈现流媒体视频和三维模型,有助于采用流媒体视频向用户呈现三维模型相关的信息,提高了交互方式的多样性,可以通过多维度信息交互,让用户更加沉静的浏览三维模型,提升用户的浏览时长,有助于满足用户更多元化的交互需求。The interactive system based on the three-dimensional model provided by the above-mentioned embodiments of the present disclosure includes a first user terminal, a second user terminal, and a server. The first user terminal presents a user interface, and the server is in communication connection with the first user terminal and the second user terminal. . Wherein: the first user terminal is configured to: in response to detecting the user's target interaction operation on the user interface, send an interaction request for the target interaction operation to the server, wherein the user interface is used to present the three-dimensional model, the three-dimensional model and the second user The user account logged in by the terminal establishes an association relationship in advance; the second user terminal is configured to: obtain the streaming video; send the streaming video to the server; the server is also configured to: send the streaming video to the first user terminal; the first user terminal It is configured to: present streaming video and three-dimensional model on the user interface. As a result, streaming media videos and 3D models can be presented on the same page of the terminal device, which helps to use streaming media videos to present information related to the 3D model to users, which improves the diversity of interaction methods. Users browse the 3D model more calmly, increase the user's browsing time, and help meet users' more diversified interactive needs.
下面参考图16来描述根据本公开实施例的电子设备。图16示出了根据本公开实施例的电子设备1600的框图。如图16所示,电子设备1600包括一个或多个处理器1611和存储器1612。The electronic device according to an embodiment of the present disclosure is described below with reference to FIG. 16. FIG. 16 shows a block diagram of an electronic device 1600 according to an embodiment of the present disclosure. As shown in FIG. 16, the electronic device 1600 includes one or more processors 1611 and a memory 1612.
处理器1611可以是中央处理单元(CPU)或者具有用于实现三维空间场景互动的能力和/或指令执行能力的其他形式的处理单元,并且可以控制电子设备1600中的其他组件以执行期望的功能。The processor 1611 may be a central processing unit (CPU) or another form of processing unit having the ability to implement three-dimensional scene interaction and/or instruction execution capabilities, and may control other components in the electronic device 1600 to perform desired functions .
存储器1612可以包括一个或多个计算机程序产品,所述计算机程序产品可以包括各种形式的计算机可读存储介质,例如易失性存储器和/或非易失性存储器。所述易失性存储器,例如,可以包括:随机存取存储器(RAM)和/或高速缓冲存储器(cache)等。所述非易失性存储器,例如,可以包括:只读存储器(ROM)、硬盘以及闪存等。在所述计算机可读存储介质上可以存储一个或多个计算机程序指令,处理器1611可以运行所述程序指令,以实现上文所述的各种方法以及/或者其他期望的功能。在所述计算机可读存储介质中还可以存储诸如输入信号、信号分量、噪声分量等各种内容。The memory 1612 may include one or more computer program products, and the computer program products may include various forms of computer-readable storage media, such as volatile memory and/or non-volatile memory. The volatile memory, for example, may include: random access memory (RAM) and/or cache memory (cache). The non-volatile memory, for example, may include: read-only memory (ROM), hard disk, flash memory, and the like. One or more computer program instructions may be stored on the computer-readable storage medium, and the processor 1611 may run the program instructions to implement the various methods described above and/or other desired functions. Various contents such as input signals, signal components, noise components, etc. can also be stored in the computer-readable storage medium.
在一个示例中,电子设备1600还可以包括:输入装置1613以及输出装置1614等,这些组件通过总线系统和/或其他形式的连接机构(未示出)互连。此外,该输入设备1613还可以包括例如键盘、鼠标等等。该输出装置1614可以向外部输出各种信息。该输出设备1614可以包括例如显示器、扬声器、打印机、以及通信网络及其所连接的远程输出设备等等。In an example, the electronic device 1600 may further include: an input device 1613, an output device 1614, etc., and these components are interconnected by a bus system and/or other forms of connection mechanisms (not shown). In addition, the input device 1613 may also include, for example, a keyboard, a mouse, and so on. The output device 1614 can output various information to the outside. The output device 1614 may include, for example, a display, a speaker, a printer, a communication network and a remote output device connected to it, and so on.
为了简化,图16中仅示出了该电子设备1600中与本公开实施例有关的组件中的一些,省略了诸如总线、输入/输出接口等等的组件。除此之外,根据具体应用情况,电子设备1600还可以包括任何其他适当的组件。除了上述方法和设备以外,本公开的实施例还可以是计算机程序产品,其包括计算机程序指令,所述计算机程序指令在被处理器运行时使得所述处理器执行根据本公开各种实施例的各种方法中的步骤。For simplification, only some of the components related to the embodiments of the present disclosure in the electronic device 1600 are shown in FIG. 16, and components such as buses, input/output interfaces, etc. are omitted. In addition, according to specific application conditions, the electronic device 1600 may also include any other appropriate components. In addition to the above-mentioned methods and devices, the embodiments of the present disclosure may also be computer program products, which include computer program instructions that, when run by a processor, cause the processor to perform operations according to various embodiments of the present disclosure. Steps in various methods.
所述计算机程序产品可以以一种或多种程序设计语言的任意组合来编写用于执行本公开实施例操作的程序代码,所述程序设计语言包括面向对象的程序设计语言,诸如Java、C++等,还包括常规的过程式程序设计语言,诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算设备上执行、部分地在用户设备上执行、作为一个独立的软件包执行、部分在用户计算设备上部分在远程计算设备上执行、或者完全在远程计算设备或服务器上执行。The computer program product may use any combination of one or more programming languages to write program codes for performing the operations of the embodiments of the present disclosure. The programming languages include object-oriented programming languages, such as Java, C++, etc. , Also includes conventional procedural programming languages, such as "C" language or similar programming languages. The program code can be executed entirely on the user's computing device, partly on the user's device, executed as an independent software package, partly on the user's computing device and partly executed on the remote computing device, or entirely on the remote computing device or server Executed on.
此外,本公开的实施例还可以是计算机可读存储介质,其上存储有计算机程序指令,所述计算机程序指令在被处理器运行时使得所述处理器执行根据本公开各种实施例的各种方法中的步骤。In addition, the embodiments of the present disclosure may also be a computer-readable storage medium on which computer program instructions are stored. When the computer program instructions are executed by a processor, the processor executes each of the various embodiments of the present disclosure. Steps in a method.
所述计算机可读存储介质可以采用一个或多个可读介质的任意组合。可读介质可以是可读信号介质或者可读存储介质。可读存储介质例如可以包括但不限于电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。可读存储介质的更具体的例子(非穷举 的列举)可以包括:具有一个或者多个导线的电连接、便携式盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。The computer-readable storage medium may adopt any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. The readable storage medium may include, but is not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the above, for example. More specific examples (non-exhaustive list) of readable storage media may include: electrical connections with one or more wires, portable disks, hard disks, random access memory (RAM), read-only memory (ROM), Erasable programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
以上结合具体实施例描述了本公开的基本原理,但是,需要指出的是,在本公开中提及的优点、优势、效果等仅是示例而非限制,不能认为这些优点、优势以及效果等是本公开的各个实施例必须具备的。另外,上述公开的具体细节仅是为了示例的作用和便于理解的作用,而非限制,上述细节并不限制本公开为必须采用上述具体的细节来实现。The above describes the basic principles of the present disclosure in conjunction with specific embodiments. However, it should be pointed out that the advantages, advantages, effects, etc. mentioned in the present disclosure are only examples and not limitations. These advantages, advantages, effects, etc. cannot be considered as Required for each embodiment of the present disclosure. In addition, the specific details of the foregoing disclosure are only for illustrative purposes and easy-to-understand functions, rather than limitations, and the foregoing details do not limit the present disclosure to the foregoing specific details for implementation.
本说明书中各个实施例均采用递进的方式描述,每个实施例重点说明的都是与其它实施例的不同之处,各个实施例之间相同或相似的部分相互参见即可。对于系统实施例而言,由于其与方法实施例基本对应,所以描述的比较简单,相关之处参见方法实施例的部分说明即可。The various embodiments in this specification are described in a progressive manner, and each embodiment focuses on the differences from other embodiments, and the same or similar parts between the various embodiments can be referred to each other. As for the system embodiment, since it basically corresponds to the method embodiment, the description is relatively simple, and the relevant part can refer to the part of the description of the method embodiment.
本公开中涉及的器件、装置、设备、系统的方框图仅作为例示性的例子并且不意图要求或暗示必须按照方框图示出的方式进行连接、布置、配置。如本领域技术人员将认识到的,可以按任意方式连接、布置、配置这些器件、装置、设备以及系统。诸如“包括”、“包含、“具有”等等的词语是开放性词汇,指“包括但不限于”,且可与其互换使用。这里所使用的词汇“或”和“和”指词汇“和/或”,且可与其互换使用,除非上下文明确指示不是如此。这里所使用的词汇“诸如”指词组“诸如但不限于”,且可与其互换使用。The block diagrams of the devices, devices, equipment, and systems involved in the present disclosure are merely illustrative examples and are not intended to require or imply that they must be connected, arranged, and configured in the manner shown in the block diagrams. As those skilled in the art will recognize, these devices, devices, devices, and systems can be connected, arranged, and configured in any manner. Words such as "include", "include, "have", etc. are open words that refer to "including but not limited to" and can be used interchangeably. The words "or" and "and" used herein refer to the words " And/or", and can be used interchangeably, unless the context clearly indicates otherwise. The term "such as" used herein refers to the phrase "such as but not limited to" and can be used interchangeably with it.
可能以许多方式来实现本公开的方法和装置。例如,可通过软件、硬件、固件或者软件、硬件、固件的任何组合来实现本公开的方法和装置。用于所述方法的步骤的上述顺序仅是为了进行说明,本公开的方法的步骤不限于以上具体描述的顺序,除非以其它方式特别说明。此外,在一些实施例中,还可将本公开实施为记录在记录介质中的程序,这些程序包括用于实现根据本公开实施例的方法的机器可读指令。因而,本公开还覆盖存储用于执行根据本公开实施例的方法的程序的记录介质。The method and apparatus of the present disclosure may be implemented in many ways. For example, the method and apparatus of the present disclosure can be implemented by software, hardware, firmware or any combination of software, hardware, and firmware. The above-mentioned order of the steps for the method is for illustration only, and the steps of the method of the present disclosure are not limited to the order specifically described above, unless specifically stated otherwise. In addition, in some embodiments, the present disclosure can also be implemented as programs recorded in a recording medium, and these programs include machine-readable instructions for implementing methods according to embodiments of the present disclosure. Thus, the present disclosure also covers a recording medium storing a program for executing a method according to an embodiment of the present disclosure.
还需要指出的是,在本公开的装置、设备和方法中,各部件或各步骤是可以分解和/或重新组合的。这些分解和/或重新组合应视为本公开的等效方案。It should also be pointed out that in the device, equipment and method of the present disclosure, each component or each step can be decomposed and/or recombined. These decomposition and/or recombination should be regarded as equivalent solutions of the present disclosure.
提供所公开的方面的以上描述,以使本领域的任何技术人员能够做出或者使用本公开。对这些方面的各种修改等对于本领域技术人员而言,是非常显而易见的,并且在此定义的一般原理可以应用于其他方面,而不脱离本公开的范围。因此,本公开不意图被限制到在此示出的方面,而是按照与在此公开的原理和新颖的特征一致的最宽范围。The above description of the disclosed aspects is provided to enable any person skilled in the art to make or use the present disclosure. Various modifications to these aspects are very obvious to those skilled in the art, and the general principles defined herein can be applied to other aspects without departing from the scope of the present disclosure. Therefore, the present disclosure is not intended to be limited to the aspects shown here, but in accordance with the widest scope consistent with the principles and novel features disclosed herein.
为了例示和描述的目的已经给出了以上描述。此外,此描述不意图将本公开的实施例限制到在此公开的形式中。尽管以上已经讨论了多个示例方面以及实施例,但是本领域技术人员将 认识到其某些变型、修改、改变、添加和子组合。The above description has been given for the purposes of illustration and description. In addition, this description is not intended to limit the embodiments of the present disclosure to the form disclosed herein. Although a number of example aspects and embodiments have been discussed above, those skilled in the art will recognize certain variations, modifications, changes, additions, and subcombinations thereof.

Claims (35)

  1. 一种用于实现三维空间场景互动的方法,包括:A method for realizing the interaction of three-dimensional space scenes, including:
    响应于检测到在三维空间场景中设置足迹信息的用户操作,确定所述用户在所述三维空间场景中的当前视角所对应的当前视图中的第一像素点;In response to detecting a user operation of setting footprint information in the three-dimensional space scene, determining the first pixel in the current view corresponding to the current perspective of the user in the three-dimensional space scene;
    确定所述第一像素点对应的三维模型;Determining a three-dimensional model corresponding to the first pixel;
    确定所述用户的足迹信息在所述三维模型中的位置,其中所述足迹信息用于在所述三维空间场景被浏览时显示;以及Determining the position of the user's footprint information in the three-dimensional model, where the footprint information is used to display when the three-dimensional space scene is browsed; and
    在所述位置处设置所述用户的足迹信息。Set the user's footprint information at the location.
  2. 根据权利要求1所述的方法,其中,所述足迹信息包括由以下各项所组成的组中的至少一项:The method according to claim 1, wherein the footprint information includes at least one item in the group consisting of:
    文本、图片、音频、视频以及三维模型。Text, pictures, audio, video and 3D models.
  3. 根据权利要求1或2所述的方法,其中,所述确定所述用户在所述三维空间场景中的当前视角所对应的当前视图中的第一像素点,包括:The method according to claim 1 or 2, wherein the determining the first pixel in the current view corresponding to the current perspective of the user in the three-dimensional space scene comprises:
    确定所述当前视图的中心像素点作为所述第一像素点。Determine the center pixel of the current view as the first pixel.
  4. 根据权利要求3所述的方法,其中,所述确定所述第一像素点对应的三维模型,包括:The method according to claim 3, wherein the determining the three-dimensional model corresponding to the first pixel point comprises:
    确定针对所述第一像素点是否设置有三维模型;Determining whether a three-dimensional model is set for the first pixel;
    响应于针对所述第一像素点设置有三维模型的确定,将所述针对所述第一像素点设置的三维模型作为所述第一像素点对应的三维模型;以及In response to determining that a three-dimensional model is set for the first pixel, using the three-dimensional model set for the first pixel as the three-dimensional model corresponding to the first pixel; and
    响应于针对所述第一像素点未设置有三维模型的确定,将针对所述当前视图中的其他像素点设置的三维模型作为所述第一像素点对应的三维模型。In response to the determination that the three-dimensional model is not set for the first pixel, the three-dimensional model set for the other pixels in the current view is used as the three-dimensional model corresponding to the first pixel.
  5. 根据权利要求4所述的方法,其中,所述将针对所述当前视图中的其他像素点设置的三维模型作为所述第一像素点对应的三维模型,包括:The method according to claim 4, wherein the using a three-dimensional model set for other pixels in the current view as a three-dimensional model corresponding to the first pixel includes:
    以所述第一像素点为起点,根据预设检查规则,对所述三维空间场景中的当前视角所对应的当前视图中的其他像素点进行检查;Using the first pixel as a starting point, and according to a preset inspection rule, inspect other pixels in the current view corresponding to the current perspective in the three-dimensional space scene;
    响应于确定检查到设置有三维模型的像素点,将所述第一像素点更新为所述设置有三维模型的像素点;In response to determining that the pixel point provided with the three-dimensional model is detected, updating the first pixel point to the pixel point provided with the three-dimensional model;
    获得所述第一像素点对应的三维模型;和Obtaining a three-dimensional model corresponding to the first pixel; and
    停止所述检查。Stop the inspection.
  6. 根据权利要求1或2所述的方法,其中,所述确定所述用户在所述三维空间场景中的当前视角所对应的当前视图中的第一像素点,包括:The method according to claim 1 or 2, wherein the determining the first pixel in the current view corresponding to the current perspective of the user in the three-dimensional space scene comprises:
    响应于所述用户在所述当前视图中设置所述足迹信息的目标位置的操作,确定所述足迹信息的目标位置对应的当前视图中的像素点,以作为所述第一像素点。In response to the user's operation of setting the target position of the footprint information in the current view, a pixel point in the current view corresponding to the target position of the footprint information is determined as the first pixel point.
  7. 根据权利要求6所述的方法,其中,所述确定所述第一像素点对应的三维模型,包括:The method according to claim 6, wherein the determining the three-dimensional model corresponding to the first pixel point comprises:
    确定针对所述第一像素点是否设置有三维模型;Determining whether a three-dimensional model is set for the first pixel;
    响应于针对所述第一像素点设置有三维模型的确定,将所述针对所述第一像素点设置的三维模型作为所述第一像素点对应的三维模型;In response to determining that a three-dimensional model is set for the first pixel point, using the three-dimensional model set for the first pixel point as a three-dimensional model corresponding to the first pixel point;
    响应于针对所述第一像素点未设置有三维模型的确定,输出更新所述足迹信息的目标位置的提示信息;In response to the determination that a three-dimensional model is not set for the first pixel point, outputting prompt information for updating the target position of the footprint information;
    响应于针对更新后的所述足迹信息的目标位置对应的当前视图中的像素点设置有三维模型的确定,将该设置有三维模型的像素点作为第一像素点;以及In response to determining that the pixel in the current view corresponding to the updated target position of the footprint information is set with a three-dimensional model, use the pixel set with the three-dimensional model as the first pixel; and
    获得所述第一像素点对应的三维模型。A three-dimensional model corresponding to the first pixel point is obtained.
  8. 根据权利要求5或7所述的方法,其中,所述确定所述用户的足迹信息在所述三维模型中的位置,包括:The method according to claim 5 or 7, wherein the determining the position of the user's footprint information in the three-dimensional model comprises:
    获取所述第一像素点在所述三维模型中的位置以作为所述用户的足迹信息在所述三维模型中的位置。Acquiring the position of the first pixel in the three-dimensional model as the position of the user's footprint information in the three-dimensional model.
  9. 根据权利要求1至8中任一项所述的方法,还包括:The method according to any one of claims 1 to 8, further comprising:
    对于浏览所述三维空间场景的任一浏览用户,确定该浏览用户在所述三维空间场景中的当前视角所对应的足迹区域;For any browsing user who browses the three-dimensional space scene, determine the footprint area corresponding to the current perspective of the browsing user in the three-dimensional space scene;
    确定所述三维模型中的属于所述足迹区域的足迹信息;以及Determining the footprint information belonging to the footprint area in the three-dimensional model; and
    在该浏览用户在所述三维空间场景中的当前视角所对应的当前视图中,显示所述属于所述足迹区域的足迹信息。In the current view corresponding to the current perspective of the browsing user in the three-dimensional space scene, the footprint information belonging to the footprint area is displayed.
  10. 根据权利要求9所述的方法,其中,所述确定该浏览用户在所述三维空间场景中的当前视角所对应的足迹区域,包括:The method according to claim 9, wherein the determining the footprint area corresponding to the current perspective of the browsing user in the three-dimensional space scene comprises:
    确定该浏览用户在所述三维空间场景中的当前视角所对应的当前视图的中心像素点;以及Determining the center pixel of the current view corresponding to the current viewing angle of the browsing user in the three-dimensional space scene; and
    以所述中心像素点为圆心,以预定长度为半径,确定所述当前视图中的足迹区域。The footprint area in the current view is determined by taking the central pixel as the center of the circle and the predetermined length as the radius.
  11. 根据权利要求9或10所述的方法,其中,所述显示所述属于所述足迹区域的足迹信息,包括:The method according to claim 9 or 10, wherein the displaying the footprint information belonging to the footprint area comprises:
    响应于确定所述属于所述足迹区域的足迹信息具有不同位置的多个足迹信息,根据所述多个足迹信息分别在所述当前视图中的图像位置,在所述当前视图中显示所述多个足迹信息;以及In response to determining that the footprint information belonging to the footprint area has multiple pieces of footprint information with different positions, displaying the multiple pieces of information in the current view according to the image positions of the multiple pieces of footprint information respectively in the current view. Footprint information; and
    响应于确定所述属于所述足迹区域的足迹信息包括相同位置的不同足迹信息,在所述当前视图中为所述不同足迹信息分配不同的图像位置,并根据所述分配的图像位置,在所述当前视图中显示所述不同足迹信息。In response to determining that the footprint information belonging to the footprint area includes different footprint information at the same location, different image positions are assigned to the different footprint information in the current view, and according to the assigned image positions, The different footprint information is displayed in the current view.
  12. 根据权利要求9至11中任一项所述的方法,还包括:The method according to any one of claims 9 to 11, further comprising:
    确定所述三维模型中的不属于所述足迹区域或当前视图的至少一个足迹信息;以及Determining at least one piece of footprint information in the three-dimensional model that does not belong to the footprint area or the current view; and
    以弹幕的形式,在该浏览用户在所述三维空间场景中的当前视角所对应的当前视图中,显示所述至少一个足迹信息。In the form of a bullet screen, the at least one piece of footprint information is displayed in the current view corresponding to the current perspective of the browsing user in the three-dimensional space scene.
  13. 一种基于三维模型的交互方法,包括:An interactive method based on a three-dimensional model, including:
    在呈现有用户界面的第一用户终端处:At the first user terminal presented with the user interface:
    响应于检测到用户针对所述用户界面的目标交互操作,向为所述用户界面提供页面数据的服务器发送针对所述目标交互操作的交互请求,其中,所述用户界面用于呈现三维模型,所述三维模型与第二用户终端登录的用户账号预先建立关联关系;In response to detecting a user's target interaction operation on the user interface, sending an interaction request for the target interaction operation to a server that provides page data for the user interface, wherein the user interface is used to present a three-dimensional model, so The three-dimensional model has a pre-established association relationship with the user account logged in by the second user terminal;
    接收所述服务器从所述第二用户终端获取的流媒体视频;以及Receiving the streaming video obtained by the server from the second user terminal; and
    在所述用户界面上呈现所述流媒体视频和所述三维模型。The streaming video and the three-dimensional model are presented on the user interface.
  14. 根据权利要求13所述的方法,其中,所述接收所述服务器从所述第二用户终端获取的流媒体视频,包括:The method according to claim 13, wherein the receiving the streaming video obtained by the server from the second user terminal comprises:
    响应于所述服务器接收到所述第二用户终端针对所述交互请求发送的交互确认信息,接收所述服务器从所述第二用户终端获取的流媒体视频。In response to the server receiving the interaction confirmation information sent by the second user terminal in response to the interaction request, receiving the streaming media video obtained by the server from the second user terminal.
  15. 根据权利要求13或14所述的方法,还包括:The method according to claim 13 or 14, further comprising:
    响应于所述第一用户终端的当前网速值小于或等于预设网速阈值,基于所述流媒体视频中的各帧语音,对目标用户图像进行调整,生成不同于所述流媒体视频的新视频,其中,所述新视频表征所述目标用户图像指示的用户执行所述各帧语音指示的动作;以及In response to the current network speed value of the first user terminal being less than or equal to the preset network speed threshold, based on each frame of voice in the streaming video, the target user image is adjusted to generate an image different from the streaming video A new video, wherein the new video characterizes that the user indicated by the target user image performs the actions indicated by each frame of voice; and
    采用所述新视频替代所述流媒体视频进行呈现。The new video is used to replace the streaming video for presentation.
  16. 根据权利要求15所述的方法,还包括:The method according to claim 15, further comprising:
    基于所述流媒体视频中的图像,生成所述目标用户图像;或者Generate the target user image based on the image in the streaming video; or
    将与所述用户账号相关联的用户图像确定为所述目标用户图像。The user image associated with the user account is determined as the target user image.
  17. 根据权利要求15或16所述的方法,还包括:The method according to claim 15 or 16, further comprising:
    响应于所述用户界面呈现有所述新视频,向所述服务器发送摄像头关闭确认信息,其中,所述摄像头关闭确认信息用于确定所述第二用户终端是否关闭摄像头。In response to the new video being presented on the user interface, sending camera shutdown confirmation information to the server, where the camera shutdown confirmation information is used to determine whether the second user terminal turns off the camera.
  18. 根据权利要求13所述的方法,其中,所述接收所述服务器从所述第二用户终端获取 的流媒体视频,包括:The method according to claim 13, wherein said receiving the streaming video obtained by the server from the second user terminal comprises:
    将所述第一用户终端的当前网速值发送至所述服务器;以及Sending the current network speed value of the first user terminal to the server; and
    接收所述服务器从所述第二用户终端获取并发送的流媒体视频,该流媒体视频具有与所述当前网速值相匹配的分辨率。Receiving a streaming video that the server obtains and sends from the second user terminal, where the streaming video has a resolution matching the current network speed value.
  19. 根据权利要求13-18之一所述的方法,还包括:The method according to any one of claims 13-18, further comprising:
    接收所述服务器发送的模型调整信息,其中,所述模型调整信息指示使用所述第二用户终端的用户对呈现于所述第二用户终端的所述三维模型的调整操作,所述调整操作包括由以下各项所组成的组中的至少一项:缩放、旋转、移动和视点切换;以及Receiving model adjustment information sent by the server, where the model adjustment information indicates an adjustment operation of the user using the second user terminal on the three-dimensional model presented on the second user terminal, the adjustment operation including At least one item from the group consisting of: zooming, rotating, moving and viewpoint switching; and
    按照所述模型调整信息指示的调整操作,对所述用户界面上呈现的所述三维模型进行相同调整操作。Perform the same adjustment operation on the three-dimensional model presented on the user interface according to the adjustment operation indicated by the model adjustment information.
  20. 根据权利要求13-19之一所述的方法,其中,所述三维模型包括多个子空间场景的三维子模型,所述多个子空间场景中的相应子空间场景与预先确定的关键词集合中的相应关键词相对应,所述方法还包括:The method according to any one of claims 13-19, wherein the three-dimensional model comprises three-dimensional sub-models of a plurality of sub-space scenes, and the corresponding sub-space scenes of the plurality of sub-space scenes correspond to those in a predetermined set of keywords. Corresponding to the corresponding keywords, the method further includes:
    对所述流媒体视频中的语音进行语音识别,得到语音识别结果;以及Perform voice recognition on the voice in the streaming video to obtain a voice recognition result; and
    响应于所述语音识别结果包含所述关键词集合中的关键词的确定,在所述用户界面上呈现与所述语音识别结果包含的关键词相对应的所述多个子空间场景中的对应子空间场景的三维子模型。In response to the determination that the voice recognition result includes the keywords in the keyword set, the corresponding subspace scenes in the plurality of subspace scenes corresponding to the keywords included in the voice recognition result are presented on the user interface. The three-dimensional sub-model of the space scene.
  21. 根据权利要求1320之一所述的方法,还包括:The method according to one of claims 1320, further comprising:
    获取用户针对所述流媒体视频的反馈信息;以及Obtaining user feedback information on the streaming media video; and
    将所述反馈信息发送至所述服务器,其中,所述服务器用于将所述反馈信息与所述用户账号建立关联关系。The feedback information is sent to the server, where the server is used to establish an association relationship between the feedback information and the user account.
  22. 一种基于三维模型的交互方法,包括:An interactive method based on a three-dimensional model, including:
    在第二用户终端处:At the second user terminal:
    响应于接收到服务器发送的交互请求,获取流媒体视频,其中,所述交互请求指示第一用户终端检测到用户针对所述第一用户终端呈现的用户界面的目标交互操作,所述用户界面用于呈现三维模型,所述三维模型与所述第二用户终端登录的用户账号预先建立关联关系;以及In response to receiving the interaction request sent by the server, the streaming video is obtained, where the interaction request indicates that the first user terminal detects the user's target interaction operation on the user interface presented by the first user terminal, and the user interface uses Presenting a three-dimensional model, the three-dimensional model pre-establishes an association relationship with the user account logged in by the second user terminal; and
    向所述服务器发送所述流媒体视频,其中,所述服务器用于将所述流媒体视频发送至所述第一用户终端,以使所述第一用户终端在所述用户界面上呈现所述流媒体视频和所述三维模型。Sending the streaming video to the server, where the server is configured to send the streaming video to the first user terminal, so that the first user terminal displays the Streaming video and the three-dimensional model.
  23. 根据权利要求22所述的方法,其中,所述响应于接收到服务器发送的交互请求,获取流媒体视频,包括:The method according to claim 22, wherein said acquiring the streaming video in response to receiving an interaction request sent by the server comprises:
    响应于接收到服务器发送的交互请求,确定是否检测到用户针对所述交互请求的确认操作;以及In response to receiving an interaction request sent by the server, determining whether a user's confirmation operation for the interaction request is detected; and
    响应于检测到所述确认操作,获取流媒体视频。In response to detecting the confirmation operation, the streaming video is acquired.
  24. 根据权利要求22或23所述的方法,还包括:The method according to claim 22 or 23, further comprising:
    响应于所述第一用户终端的当前网速值小于或等于预设网速阈值,从所述服务器接收摄像头关闭确认信息;以及In response to the current network speed value of the first user terminal being less than or equal to the preset network speed threshold, receiving camera shutdown confirmation information from the server; and
    呈现所述摄像头关闭确认信息,其中,所述摄像头关闭确认信息用于确定所述第二用户终端是否关闭摄像头。Presenting the camera closing confirmation information, where the camera closing confirmation information is used to determine whether the second user terminal closes the camera.
  25. 根据权利要求22-24之一所述的方法,还包括:The method according to any one of claims 22-24, further comprising:
    响应于从所述服务器接收到用户对呈现于所述第一用户终端的所述三维模型的调整操作,按照所述模型调整信息指示的调整操作,对所述第二用户终端呈现的所述三维模型进行相同调整操作,其中,所述调整操作包括由以下各项所组成的组中的至少一项:缩放、旋转、移动和视点切换。In response to receiving from the server the user's adjustment operation on the three-dimensional model presented on the first user terminal, according to the adjustment operation indicated by the model adjustment information, perform the adjustment on the three-dimensional model presented by the second user terminal. The model performs the same adjustment operation, wherein the adjustment operation includes at least one item from the group consisting of: zooming, rotating, moving, and viewpoint switching.
  26. 根据权利要求22-25之一所述的方法,还包括:The method according to any one of claims 22-25, further comprising:
    响应于检测到用户对呈现于所述第二用户终端的所述三维模型的调整操作,向所述服务器发送指示所述调整操作的模型调整信息,以使所述服务器控制所述第一用户终端按照所述模型调整信息指示的调整操作,对所述用户界面上呈现的所述三维模型进行相同调整操作,其中,所述调整操作包括由以下各项所组成的组中的至少一项:缩放、旋转、移动和视点切换。In response to detecting the user's adjustment operation on the three-dimensional model presented on the second user terminal, sending model adjustment information indicating the adjustment operation to the server, so that the server controls the first user terminal Perform the same adjustment operation on the three-dimensional model presented on the user interface according to the adjustment operation indicated by the model adjustment information, wherein the adjustment operation includes at least one item in the group consisting of: zooming , Rotation, movement and viewpoint switching.
  27. 根据权利要求22-26之一所述的方法,其中,所述三维模型包括多个子空间场景的三维子模型,所述多个子空间场景中的相应子空间场景与预先确定的关键词集合中的相应关键词相对应,The method according to any one of claims 22-26, wherein the three-dimensional model comprises three-dimensional sub-models of a plurality of sub-space scenes, and the corresponding sub-space scenes in the plurality of sub-space scenes correspond to those in a predetermined set of keywords. Corresponding to the corresponding keywords,
    所述方法还包括:The method also includes:
    对所述第一用户终端获取的语音进行语音识别,得到语音识别结果;以及Performing voice recognition on the voice acquired by the first user terminal to obtain a voice recognition result; and
    响应于所述语音识别结果包含所述关键词集合中的关键词的确定,在所述用户界面上呈现与所述语音识别结果包含的关键词相对应的所述多个子空间场景中的对应子空间场景的三维子模型。In response to the determination that the voice recognition result includes the keywords in the keyword set, corresponding subspace scenes in the plurality of subspace scenes corresponding to the keywords included in the voice recognition result are presented on the user interface The three-dimensional sub-model of the space scene.
  28. 根据权利要求22-27之一所述的方法,还包括:The method according to any one of claims 22-27, further comprising:
    响应于接收到所述服务器发送的、使用所述第一用户终端的用户针对所述流媒体视频的反 馈信息,执行与所述反馈信息相匹配的操作。In response to receiving the feedback information for the streaming video from the user using the first user terminal sent by the server, perform an operation that matches the feedback information.
  29. 一种用于实现三维空间场景互动的装置,包括:用于执行权利要求1至12中任一项所述的方法的装置。A device for realizing interaction in a three-dimensional space scene, comprising: a device for executing the method according to any one of claims 1 to 12.
  30. 一种基于三维模型的交互装置,设置于第一用户终端,其中,所述装置包括:用于执行权利要求13至21中任一项所述的方法的装置。An interaction device based on a three-dimensional model, which is provided in a first user terminal, wherein the device comprises: a device for executing the method according to any one of claims 13 to 21.
  31. 一种基于三维模型的交互装置,设置于第二用户终端,其中,所述装置包括:用于执行权利要求22-28中任一项所述的方法的装置。An interaction device based on a three-dimensional model, which is provided in a second user terminal, wherein the device includes: a device for executing the method according to any one of claims 22-28.
  32. 一种基于三维模型的交互系统,包括:An interactive system based on a three-dimensional model, including:
    第一用户终端,用于呈现用户界面;The first user terminal is used to present a user interface;
    第二用户终端;和The second user terminal; and
    服务器,所述服务器与所述第一用户终端和所述第二用户终端通信连接,A server, the server is in communication connection with the first user terminal and the second user terminal,
    其中,所述第一用户终端被配置成:响应于检测到用户针对所述用户界面的目标交互操作,向所述服务器发送针对所述目标交互操作的交互请求,其中,所述用户界面用于呈现三维模型,所述三维模型与第二用户终端登录的用户账号预先建立关联关系;Wherein, the first user terminal is configured to: in response to detecting a user's target interaction operation for the user interface, send an interaction request for the target interaction operation to the server, wherein the user interface is used for Presenting a three-dimensional model, which has a pre-established association relationship with the user account logged in by the second user terminal;
    其中,所述第二用户终端被配置成:获取流媒体视频;并且向所述服务器发送所述流媒体视频;Wherein, the second user terminal is configured to: obtain a streaming video; and send the streaming video to the server;
    其中,所述服务器被配置成:向所述第一用户终端发送所述流媒体视频;并且Wherein, the server is configured to: send the streaming video to the first user terminal; and
    其中,所述第一用户终端被配置成:在所述用户界面上呈现所述流媒体视频和所述三维模型。Wherein, the first user terminal is configured to: present the streaming video and the three-dimensional model on the user interface.
  33. 一种非暂态性计算机可读存储介质,所述存储介质存储有计算机程序,所述计算机程序当被计算机执行时使所述计算机实现上述权利要求1-28中任一项所述的方法。A non-transitory computer-readable storage medium storing a computer program that, when executed by a computer, enables the computer to implement the method according to any one of claims 1-28.
  34. 一种电子设备,包括:An electronic device including:
    处理器;以及Processor; and
    用于存储处理器可执行指令的存储器,所述处理器可执行指令在被所述处理器执行时实现上述权利要求1-28中任一项所述的方法。A memory for storing processor-executable instructions that, when executed by the processor, implement the method according to any one of claims 1-28.
  35. 一种计算机程序产品,包括计算机程序,所述计算机程序当被计算机执行时使所述计算机实现上述权利要求1-28中任一项所述的方法。A computer program product, comprising a computer program, which when executed by a computer causes the computer to implement the method according to any one of claims 1-28.
PCT/CN2021/093628 2020-05-13 2021-05-13 Method for realizing interaction in three-dimensional space scene, apparatus and device WO2021228200A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN202010401813.7 2020-05-13
CN202010401813.7A CN111562845B (en) 2020-05-13 2020-05-13 Method, device and equipment for realizing three-dimensional space scene interaction
CN202010698810.4 2020-07-20
CN202010698810.4A CN111885398B (en) 2020-07-20 2020-07-20 Interaction method, device and system based on three-dimensional model, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
WO2021228200A1 true WO2021228200A1 (en) 2021-11-18

Family

ID=78525899

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/093628 WO2021228200A1 (en) 2020-05-13 2021-05-13 Method for realizing interaction in three-dimensional space scene, apparatus and device

Country Status (1)

Country Link
WO (1) WO2021228200A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114241132A (en) * 2021-12-16 2022-03-25 北京字跳网络技术有限公司 Scene content display control method and device, computer equipment and storage medium
CN115499641A (en) * 2022-09-20 2022-12-20 北京三月雨文化传播有限责任公司 Method for quickly constructing digital exhibition file and intelligent terminal

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107710284A (en) * 2015-06-30 2018-02-16 奇跃公司 For more effectively showing the technology of text in virtual image generation system
CN108874471A (en) * 2018-05-30 2018-11-23 链家网(北京)科技有限公司 Additional elements adding method and system between a kind of function of the source of houses
CN110531847A (en) * 2019-07-26 2019-12-03 中国人民解放军军事科学院国防科技创新研究院 A kind of novel social contact method and system based on augmented reality
CN110891167A (en) * 2019-11-30 2020-03-17 北京城市网邻信息技术有限公司 Information interaction method, first terminal and computer readable storage medium
CN110944140A (en) * 2019-11-30 2020-03-31 北京城市网邻信息技术有限公司 Remote display method, remote display system, electronic device and storage medium
CN111047717A (en) * 2019-12-24 2020-04-21 北京法之运科技有限公司 Method for carrying out character labeling on three-dimensional model
CN111562845A (en) * 2020-05-13 2020-08-21 贝壳技术有限公司 Method, device and equipment for realizing three-dimensional space scene interaction
CN111885398A (en) * 2020-07-20 2020-11-03 贝壳技术有限公司 Interaction method, device and system based on three-dimensional model

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107710284A (en) * 2015-06-30 2018-02-16 奇跃公司 For more effectively showing the technology of text in virtual image generation system
CN108874471A (en) * 2018-05-30 2018-11-23 链家网(北京)科技有限公司 Additional elements adding method and system between a kind of function of the source of houses
CN110531847A (en) * 2019-07-26 2019-12-03 中国人民解放军军事科学院国防科技创新研究院 A kind of novel social contact method and system based on augmented reality
CN110891167A (en) * 2019-11-30 2020-03-17 北京城市网邻信息技术有限公司 Information interaction method, first terminal and computer readable storage medium
CN110944140A (en) * 2019-11-30 2020-03-31 北京城市网邻信息技术有限公司 Remote display method, remote display system, electronic device and storage medium
CN111047717A (en) * 2019-12-24 2020-04-21 北京法之运科技有限公司 Method for carrying out character labeling on three-dimensional model
CN111562845A (en) * 2020-05-13 2020-08-21 贝壳技术有限公司 Method, device and equipment for realizing three-dimensional space scene interaction
CN111885398A (en) * 2020-07-20 2020-11-03 贝壳技术有限公司 Interaction method, device and system based on three-dimensional model

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114241132A (en) * 2021-12-16 2022-03-25 北京字跳网络技术有限公司 Scene content display control method and device, computer equipment and storage medium
CN114241132B (en) * 2021-12-16 2023-07-21 北京字跳网络技术有限公司 Scene content display control method and device, computer equipment and storage medium
CN115499641A (en) * 2022-09-20 2022-12-20 北京三月雨文化传播有限责任公司 Method for quickly constructing digital exhibition file and intelligent terminal
CN115499641B (en) * 2022-09-20 2023-09-12 广东鸿威国际会展集团有限公司 Method for quickly constructing digital exhibition file and intelligent terminal

Similar Documents

Publication Publication Date Title
WO2021109652A1 (en) Method and apparatus for giving character virtual gift, device, and storage medium
US11899900B2 (en) Augmented reality computing environments—immersive media browser
US10134364B2 (en) Prioritized display of visual content in computer presentations
CN111178191B (en) Information playing method and device, computer readable storage medium and electronic equipment
JP5901151B2 (en) How to select objects in a virtual environment
KR20220115824A (en) Matching content to a spatial 3d environment
WO2018152455A1 (en) System and method for creating a collaborative virtual session
WO2022116751A1 (en) Interaction method and apparatus, and terminal, server and storage medium
US20180160194A1 (en) Methods, systems, and media for enhancing two-dimensional video content items with spherical video content
WO2021228200A1 (en) Method for realizing interaction in three-dimensional space scene, apparatus and device
US11908056B2 (en) Sentiment-based interactive avatar system for sign language
CN112596694B (en) Method and device for processing house source information
WO2022095757A1 (en) Image rendering method and apparatus
KR20190047144A (en) Interactive video generation
US20230409632A1 (en) Systems and methods for using conjunctions in a voice input to cause a search application to wait for additional inputs
US20230018502A1 (en) Display apparatus and method for person recognition and presentation
US20220319063A1 (en) Method and apparatus for video conferencing
CN112051956A (en) House source interaction method and device
TWI570639B (en) Systems and methods for building virtual communities
CN111562845B (en) Method, device and equipment for realizing three-dimensional space scene interaction
CN111885398B (en) Interaction method, device and system based on three-dimensional model, electronic equipment and storage medium
CN116017082A (en) Information processing method and electronic equipment
WO2022205001A1 (en) Information exchange method, computer-readable storage medium, and communication terminal
CN116762333A (en) Superimposing images of conference call participants and shared documents

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21805221

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21805221

Country of ref document: EP

Kind code of ref document: A1