WO2023236815A1 - Three-dimensional model transmission method and apparatus, and storage medium and program product - Google Patents

Three-dimensional model transmission method and apparatus, and storage medium and program product Download PDF

Info

Publication number
WO2023236815A1
WO2023236815A1 PCT/CN2023/097180 CN2023097180W WO2023236815A1 WO 2023236815 A1 WO2023236815 A1 WO 2023236815A1 CN 2023097180 W CN2023097180 W CN 2023097180W WO 2023236815 A1 WO2023236815 A1 WO 2023236815A1
Authority
WO
WIPO (PCT)
Prior art keywords
model
dimensional
dimensional model
physical
dimensional scene
Prior art date
Application number
PCT/CN2023/097180
Other languages
French (fr)
Chinese (zh)
Inventor
王志强
Original Assignee
中兴通讯股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中兴通讯股份有限公司 filed Critical 中兴通讯股份有限公司
Publication of WO2023236815A1 publication Critical patent/WO2023236815A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/194Transmission of image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working

Definitions

  • Embodiments of the present application relate to but are not limited to the field of communication technology, and in particular, to a three-dimensional model transmission method and its device, storage medium, and program product.
  • Embodiments of the present application provide a three-dimensional model transmission method and its device, storage media, and program products.
  • embodiments of the present application provide a three-dimensional model transmission method, including: acquiring multiple video images from the design end, and the video images are processed by the design end by performing frame extraction on the video stream of the augmented reality video call.
  • the augmented reality video call is established by the design terminal and the client; perform segmentation processing on a plurality of the video images to obtain a set of physical images and a set of environmental images; perform modeling processing on the set of physical images, Obtain the physical three-dimensional model; perform modeling processing on the environmental image collection to obtain the environmental three-dimensional model; generate a first three-dimensional scene model according to the physical three-dimensional model and the environmental three-dimensional model; send the first three-dimensional scene model to the client.
  • embodiments of the present application provide a three-dimensional model transmission method, which includes: establishing an augmented reality video call with the design end; receiving a first three-dimensional scene model sent by the server, and the first three-dimensional scene model is sent by the server
  • the terminal is obtained based on the physical three-dimensional model and the environmental three-dimensional model.
  • the physical three-dimensional model is obtained by the server side by modeling the physical object image collection.
  • the environmental three-dimensional model is obtained by the server side by modeling the environmental image collection.
  • the physical image set and the environment image set are obtained by the server side by segmenting multiple video images, and the video images are obtained by the design end through an augmented reality video call.
  • the video stream is obtained by performing frame extraction processing.
  • embodiments of the present application also provide a three-dimensional model transmission device, including: a memory, a processor, and a computer program stored in the memory and executable on the processor.
  • a processor executes the computer program Implement the three-dimensional model transfer method as described above.
  • embodiments of the present application also provide a computer-readable storage medium that stores computer-executable instructions, and the computer-executable instructions are used to execute the three-dimensional model transmission method as described above.
  • embodiments of the present application further provide a computer program product, which includes a computer program or computer instructions.
  • the computer program or computer instructions are stored in a computer-readable storage medium.
  • the processor of the computer device obtains the information from the computer program or computer instructions.
  • the computer-readable storage medium reads the computer program or the computer instructions, and the processor executes the computer program or the computer instructions, so that the computer device performs the three-dimensional model transmission method as described above.
  • Figure 1 is a flow chart of a three-dimensional model transmission method provided by an embodiment of the present application.
  • Figure 2 is a flow chart of a method in step S150 in Figure 1;
  • Figure 3 is a flow chart of a three-dimensional model transmission method provided by another embodiment of the present application.
  • Figure 4 is a flow chart of a method in step S130 in Figure 1;
  • Figure 5 is a flow chart of a method in step S140 in Figure 1;
  • Figure 6 is a flow chart of a three-dimensional model transmission method provided by another embodiment of the present application.
  • Figure 7 is a flow chart of a three-dimensional model transmission method provided by another embodiment of the present application.
  • Figure 8 is a flow chart of a three-dimensional model transmission method provided by another embodiment of the present application.
  • Figure 9 is a flow chart of a three-dimensional model transmission method provided by an embodiment of the present application.
  • Figure 10 is a schematic structural diagram of a three-dimensional model transmission device provided by an embodiment of the present application.
  • Figure 11 is a schematic structural diagram of a three-dimensional model transmission device provided by another embodiment of the present application.
  • This application provides a three-dimensional model transmission method and its device, storage media, and program products.
  • the server side obtains multiple video images from the design side.
  • the video images are processed by the design side by extracting frames from the video stream of the augmented reality video call.
  • the augmented reality video call is established between the design end and the client, and then multiple video images are segmented to obtain a collection of physical images and a collection of environmental images, and then the collection of physical images is modeled to obtain a three-dimensional model of the physical object, and
  • the environment image collection is modeled and processed to obtain an environment three-dimensional model.
  • the first three-dimensional scene model is generated based on the physical three-dimensional model and the environment three-dimensional model.
  • the first three-dimensional scene model is sent to the client.
  • the design end performs frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images.
  • the server end processes the multiple video images to finally obtain the first three-dimensional scene. model, and sends the first three-dimensional scene model to the client. Therefore, the embodiment of the present application can realize the transmission of the three-dimensional model in the augmented reality scene.
  • Figure 1 is a flow chart of a three-dimensional model transmission method provided by an embodiment of the present application.
  • the three-dimensional model transmission method may include but is not limited to step S110, step S120, step S130, step S140, step S150 and step S160.
  • Step S110 Obtain multiple video images from the design end.
  • the video image is obtained by the design end by performing frame extraction processing on the video stream of the augmented reality video call, and the augmented reality video call is established by the design end and the client.
  • the design end can intercept a clear frame of image from the video, or it can intercept multiple image sets, that is, the design end can intercept the video stream of the augmented reality video call. Perform frame extraction processing to obtain multiple video images.
  • the design end sends a video conference invitation to the client.
  • the design end and the client can successfully establish an augmented reality video call after server-side authentication and confirmation; or, the client sends a video conference invitation to the design end.
  • the design side and the client can successfully establish an augmented reality video call.
  • audio and video switching can be supported.
  • the client can switch to the playback interface of application A and display the augmented reality video. Set the call as a background program so that the augmented reality video call runs in the background.
  • Application A can be audio or video. frequency; for another example, when the client is having an augmented reality video call with the designer, the client can switch the video call to a voice call, and there are no specific restrictions here.
  • the video stream of the augmented reality video call can adopt YUV encoding and H.264 video protocol, where YUV is a color encoding format, "Y” represents brightness, that is, grayscale value; "U” and “V” represents chroma, "U” and “V” can both describe image color and saturation, and can be used to specify the color of pixels.
  • the H.264 video protocol is a digital video compression encoding standard.
  • the video stream of the augmented reality video call can be a video stream on the client side, a video stream on the design side, or a video stream on the client side and a video stream on the user side. It is obtained based on the actual situation and there are no specific restrictions here.
  • the design end can extract frames from the video stream on the design end to obtain multiple video images; or, after the design end and the client establish an augmented reality video call, , the design end can perform frame extraction processing on the video stream on the client side to obtain multiple video images; or, after the design end and the client establish an augmented reality video call, the design end can process the video stream on the design end side and the client side respectively.
  • the video stream on the client side undergoes frame extraction processing to obtain multiple video images.
  • Step S120 Perform segmentation processing on multiple video images to obtain a set of physical images and a set of environmental images.
  • the physical object image set includes multiple physical object images
  • the environmental image set includes multiple environmental images
  • edge contour scanning may be used to segment multiple video images to obtain a set of physical images and a set of environmental images; or , the deep learning algorithm can be used to identify and process the physical object images and environmental images in multiple video images, and the segmented areas corresponding to the recognized physical images and the environmental images corresponding to the multiple video images are determined respectively according to the composition rules.
  • segmentation area and then obtain the physical image set and the environmental image set based on the segmentation area corresponding to the physical image and the segmentation area corresponding to the environmental image, wherein the composition rule may include dividing the physical image or the environmental image in the segmentation area.
  • Step S130 Perform modeling processing on the collection of real object images to obtain a three-dimensional model of the real object.
  • Step S140 Perform modeling processing on the environment image collection to obtain a three-dimensional environment model.
  • Step S150 Generate a first three-dimensional scene model based on the physical three-dimensional model and the environmental three-dimensional model.
  • Step S160 Send the first three-dimensional scene model to the client.
  • the server can send the first three-dimensional scene model to the design end, which is not specifically limited here.
  • the server side obtains multiple video images from the design side, and the video images are extracted from the video stream of the augmented reality video call by the design side. Obtained through processing, the augmented reality video call is established between the design end and the client, and then multiple video images are segmented to obtain a collection of physical images and a collection of environmental images, and then the collection of physical images is modeled to obtain a three-dimensional model of the physical object.
  • the design end performs frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images.
  • the server end processes the multiple video images to finally obtain the first three-dimensional scene model, and sends the first three-dimensional scene model to the client. Therefore, the embodiment of the present application can construct a three-dimensional model in an augmented reality scene and realize the transmission of the three-dimensional model.
  • the design end, the server end and the client can all perform three-dimensional rendering processing on the first three-dimensional scene model, which is not specifically limited here.
  • the client performs three-dimensional rendering processing on the first three-dimensional scene model sent by the server to obtain the rendered first three-dimensional scene model, and then displays the rendered first three-dimensional scene model; or, when the server sends the first three-dimensional scene model, the client performs three-dimensional rendering processing on the first three-dimensional scene model sent by the server.
  • a three-dimensional scene model is sent to the design end, and the design end performs three-dimensional rendering processing on the first three-dimensional scene model sent by the server end to obtain the rendered first three-dimensional scene model, and then displays the rendered first three-dimensional scene model;
  • the server first performs three-dimensional rendering processing on the first three-dimensional scene model to obtain the rendered first three-dimensional scene model, and then sends the rendered first three-dimensional scene model to the design end or the client.
  • the first three-dimensional scene model can be subjected to three-dimensional rendering processing multiple times, and whether it is necessary to perform three-dimensional rendering processing on the first three-dimensional scene model multiple times can be determined according to the model scale of the first three-dimensional scene model.
  • the first three-dimensional scene model simulates an amusement park
  • the first three-dimensional scene model can be subjected to three-dimensional rendering processing multiple times. That is, the server first performs three-dimensional rendering processing on the first three-dimensional scene model to obtain the rendered first three-dimensional scene model.
  • the scene model is then sent to the design end (or client), and the design end (or client) performs three-dimensional rendering processing again on the rendered first three-dimensional scene model sent by the server.
  • the re-rendered first three-dimensional scene model obtains the re-rendered first three-dimensional scene model, and display the re-rendered first three-dimensional scene model; or, when the first three-dimensional scene model simulates a bedroom, then only the first three-dimensional scene model can be A three-dimensional rendering process is performed once, and the first three-dimensional scene model can be three-dimensionally rendered by the design end, the server end, or the client, and there are no specific restrictions here.
  • 3D rendering is essentially a conversion process of computer processing of 3D models into 2D images.
  • Rendering involves scan line rendering, ray tracing, photon mapping and other technologies, which can simulate the interaction between light and various substances (such as materials and surface textures). Therefore, rendering requires the support of 3D plug-ins, 3D software and hardware.
  • step S150 is further described.
  • This step S150 may include but is not limited to step S210, step S220 and step S230.
  • Step S210 Obtain a preset three-dimensional model library, which includes an environment model library.
  • the environment model library may include multiple preset environment models.
  • the multiple preset environment models may be created by model makers using three-dimensional modeling tools (such as 3D Studio MAX and other modeling tools, where 3D Studio MAX is a three-dimensional animation rendering and production software based on a computer operating system). It is obtained by modifying the existing model data and using the modified model data for modeling processing. It can also be an existing model, which will not be done here. Specific restrictions.
  • the preset three-dimensional model library may include a target three-dimensional scene model, model-related data, model basic components, model material materials, etc., where the model-related data includes model scale, location information of the real scene simulated by the model, etc., No specific restrictions are made here.
  • the server determines the target 3D scene corresponding to the mountain peak from the preset 3D model library based on the location information. model, and sends the target three-dimensional scene model to the design end.
  • the design end plans the route of the mountain peak based on the target three-dimensional scene model and marks the iconic attractions, shops, restrooms, etc. on the mountain route, and combines the route planning and The relevant information of the mark is pushed to the AR client, thereby improving the user experience, in which the AR device is the client.
  • Step S220 Determine the target environment three-dimensional model from the environment model library according to the physical three-dimensional model and the environment three-dimensional model.
  • the physical three-dimensional model and the environmental three-dimensional model can be matched with the preset environmental model in the environmental model library to determine the target environmental three-dimensional model, thereby improving modeling efficiency and shortening modeling time.
  • Step S230 Generate a first three-dimensional scene model based on the physical three-dimensional model and the target environment three-dimensional model.
  • the server side obtains a preset three-dimensional model library, where the three-dimensional model library includes an environment model library, and then based on the physical three-dimensional model and the environmental three-dimensional
  • the model determines the target environment three-dimensional model from the environment model library, and finally generates the first three-dimensional scene model based on the physical three-dimensional model and the target environment three-dimensional model. Therefore, the embodiment of the present application can obtain the target environment three-dimensional model from the environment model library based on the physical three-dimensional model and the environment three-dimensional model. Determine the three-dimensional model of the target environment to improve modeling efficiency and shorten modeling time.
  • the three-dimensional model transmission method may also include but is not limited to step S310, step S320, step S330, step S340 and step S350.
  • Step S310 Receive the annotation information sent by the client.
  • the annotation information may include coordinate information, text information, annotated models, etc.
  • the annotation information is obtained by the client performing virtual annotation processing on the first three-dimensional scene model. For example, when the first three-dimensional scene model simulates a living room, and a sofa is placed in the living room at the coordinates (8, 4, 0), if it is necessary to change the sofa to the position at the coordinates (4, 4, 0) If placed, the coordinate information may be to modify the physical three-dimensional model at the coordinates of (8, 4, 0) to the position of the coordinates of (4, 4, 0).
  • Step S320 Determine the labeling area in the first three-dimensional scene model according to the labeling information.
  • the labeled area can be the area before modification, the area after modification, or the area before modification and the area after modification.
  • the labeling information is the coordinates (8, 4, 0)
  • the physical 3D model at is modified to If the coordinates are (4, 4, 0), then the marked area can be the area with coordinates (8, 4, 0) or the area with coordinates (4, 4, 0). There are no specific restrictions here. .
  • Step S330 Nest the annotation information on the annotation area to obtain a second three-dimensional scene model.
  • Step S340 Overlay the first three-dimensional scene model and the second three-dimensional scene model to obtain a third three-dimensional scene model.
  • Step S350 Send the third three-dimensional scene model to the client.
  • the server side can send the third three-dimensional scene model to the design side, which is not specifically limited here.
  • the server receives the annotation information sent by the client, determines the annotation area in the third three-dimensional scene model according to the annotation information, and then determines the annotation area in the annotation area.
  • the annotation information is nested on the top to obtain the second three-dimensional scene model.
  • the third three-dimensional scene model and the second three-dimensional scene model are superimposed to obtain the third three-dimensional scene model.
  • the third three-dimensional scene model is sent to the client.
  • the embodiment of this application can receive annotation information fed back by the client, and modify the third three-dimensional scene model based on the annotation information, reducing communication costs caused by professional barriers, achieving rapid iteration and delivery, and conducive to meeting customer needs. At the same time Improved user experience.
  • the design end, the server end and the client can all perform 3D rendering processing on the third 3D scene model, which is not specifically limited here.
  • the client performs three-dimensional rendering processing on the third three-dimensional scene model sent by the server to obtain the rendered third three-dimensional scene model, and then displays the rendered third three-dimensional scene model; or, when the server sends the third three-dimensional scene model, the client performs three-dimensional rendering processing.
  • the three-dimensional scene model is sent to the design end, and the design end performs three-dimensional rendering processing on the third three-dimensional scene model sent by the server to obtain a rendered third three-dimensional scene model, and then displays the rendered third three-dimensional scene model;
  • the server first performs three-dimensional rendering processing on the third three-dimensional scene model to obtain a rendered third three-dimensional scene model, and then sends the rendered third three-dimensional scene model to the design end or the client.
  • the third three-dimensional scene model can be subjected to three-dimensional rendering processing multiple times, and whether it is necessary to perform three-dimensional rendering processing on the third three-dimensional scene model multiple times can be determined according to the model scale of the third three-dimensional scene model.
  • the third 3D scene model can be 3D rendered multiple times. That is, the server first performs 3D rendering on the third 3D scene model and obtains the The third three-dimensional scene model is then sent to the design end (or client), and the design end (or client) responds to the rendered third three-dimensional scene model sent by the server again.
  • step S130 when the three-dimensional model library includes a physical model library, step S130 is further described.
  • This step S130 may include but is not limited to step S410, step S420, step S430 and Step S440.
  • Step S410 Analyze the collection of physical objects to obtain graphic information.
  • the graphic information may include the line outline of the physical image, the position of the physical image, the size of the physical image, the resource address of the physical image, etc., which are not specifically limited here.
  • Step S420 Obtain physical image element information from the physical model library according to the graphic information.
  • the physical model library may include multiple preset physical models.
  • the multiple preset physical models may be created by model makers using three-dimensional modeling tools (such as 3D Studio MAX and other modeling tools, where 3D Studio MAX is a three-dimensional animation rendering and production software based on a computer operating system). It is obtained by modifying the existing model data and using the modified model data for modeling processing. It can also be an existing model, which will not be done here. Specific restrictions.
  • Step S430 Use the element information of the physical image to construct a base map of the physical image.
  • the physical image base map may be the basic framework of the physical three-dimensional model, which is not specifically limited here.
  • Step S440 Perform modeling processing on the image base map of the physical object to obtain a three-dimensional model of the physical object.
  • the vase is modeled, that is, The vase is subjected to material veneer, light adjustment, texture, depression, and convex processing, etc., and finally a physical three-dimensional model of the vase is obtained.
  • the server side parses the physical image collection to obtain graphic information, and obtains the physical image element information from the physical model library according to the graphic information. Then, the physical image element information is used to construct the physical image base map, and finally the physical image base map is modeled to obtain the physical three-dimensional model. Therefore, the embodiment of the present application can obtain the physical image element information from the physical model library through the graphic information, using The physical image element information is used to construct the physical image base map to shorten the modeling time and improve the modeling efficiency.
  • step S140 is further described.
  • This step S140 may include but is not limited to step S510, step S520 and step S530.
  • Step S510 Analyze the environment image set to obtain label information.
  • the tag information may include material attribute information, color attribute information, light source attribute information, etc. of environmental appearance, which are not specifically limited here.
  • Step S520 Obtain environmental image element information from the environment model library according to the label information.
  • the environment image element information may include index information such as material attribute information, color attribute information, light source attribute information, etc. of the environment appearance, which is not specifically limited here.
  • Step S530 Use the environment image element information to perform modeling processing on the environment image set to obtain a three-dimensional environment model.
  • the environment image element information is used to model the environment image collection, which may include material veneer, light adjustment, texture, depression, bulge processing, collapse processing, etc., so that the model has a three-dimensional sense. This is not the case here. Make specific restrictions.
  • the server parses the environment image collection to obtain label information, and then obtains environmental image element information from the environment model library according to the label information. , and finally the environment image element information is used to model the environment image collection to obtain the environment three-dimensional model. Therefore, the embodiment of the present application obtains the environment image element information from the environment model library through the tag information, and uses the environment image element information to model the environment image collection. Carry out modeling processing to achieve the purpose of shortening modeling time and improving modeling efficiency.
  • the three-dimensional model transmission method may also include but is not limited to step S610 and step S620.
  • Step S610 Determine the target environment three-dimensional model from the environment model library according to the physical three-dimensional model.
  • the environment model library may include multiple preset environment models.
  • the multiple preset environment models may be created by model makers using three-dimensional modeling tools (such as 3D Studio MAX and other modeling tools, where 3D Studio MAX is a three-dimensional animation rendering and production software based on a computer operating system). It is obtained by modifying the existing model data and using the modified model data for modeling processing. It can also be an existing model, which will not be done here. Specific restrictions.
  • the physical three-dimensional model can be adapted to the preset environment model in the environment model library to determine the target environment three-dimensional model without modeling the environment image collection, which shortens the modeling time.
  • the application examples do not specifically limit this.
  • Step S620 Nest the physical three-dimensional model and the target environment three-dimensional model to generate a first three-dimensional scene model.
  • the server determines the target environment three-dimensional model from the environment model library based on the physical three-dimensional model, and finally embeds the physical three-dimensional model and the target environment three-dimensional model.
  • a set of processes is performed to generate the first three-dimensional scene model. That is to say, you can only perform modeling processing on the physical image collection to obtain the physical three-dimensional model. You can only determine the target environment three-dimensional model from the environment model library based on the physical three-dimensional model, and convert the physical three-dimensional model to the first three-dimensional scene model.
  • the form of the three-dimensional model is adapted to the preset environmental three-dimensional model in the environment model library to achieve the effect of remote design, meet the diverse needs of users, and improve user experience.
  • the client when designing the home layout of the living room, can send a collection of physical objects including sofas, refrigerators, seats, etc. to the server, and the server performs modeling processing on the physical image collection to obtain multiple physical objects.
  • the server determines the target environment three-dimensional model from the environment model library based on the multiple physical three-dimensional models, performs nesting processing on the physical three-dimensional model and the target environment three-dimensional model, and generates the first three-dimensional scene model. Therefore, the embodiment of the present application There is no need to perform modeling processing on the environment image collection, which shortens the modeling time.
  • step S150 is further described.
  • Step S150 may include but is not limited to the following steps:
  • the physical three-dimensional model and the environmental three-dimensional model are nested to generate a first three-dimensional scene model.
  • the server side can obtain multiple video images from the design side.
  • the video images are obtained by the design side by performing frame extraction processing on the video stream of the augmented reality video call.
  • the augmented reality video call is established by the design side and the client.
  • multiple video images are segmented to obtain a collection of physical images and a collection of environmental images.
  • the physical image collection is modeled to obtain a three-dimensional model of the physical object.
  • the environmental image collection is analyzed and processed to obtain graphic information, and then the physical object is
  • the three-dimensional model and the environment three-dimensional model are nested to generate the first three-dimensional scene model. Therefore, the embodiment of the present application can model the environment and physical objects in the scene where the user is located, and combine the physical three-dimensional model with the environmental three-dimensional model.
  • the first three-dimensional scene model is obtained through nesting processing, restoring the real scene and achieving the effect of remote customization.
  • the server side when the server side can obtain multiple video images from the design side, it performs segmentation processing on the multiple video images to obtain a set of physical images and a set of environmental images, and then performs modeling processing on the set of physical images to obtain a three-dimensional physical object. model, and parses the environment image collection to obtain label information, and adapts the label information to the preset environment model in the environment model library. When the label information does not match the preset environment model, the environment image collection is constructed. Model processing is performed to obtain an environment three-dimensional model, and the physical three-dimensional model and the environment three-dimensional model are nested to obtain the first three-dimensional scene model.
  • the environment three-dimensional model can be stored in the environment model library, and this application does not impose specific restrictions on this.
  • the server acquires the set of environment images sent by the client, and performs modeling processing on the set of environment images to obtain a three-dimensional model of the environment. Then, the server acquires multiple video images from the design end, and the video images are generated by the design end. It is obtained by extracting frames from the video stream of the augmented reality video call. The augmented reality video call is established by the design end and the client. Then multiple video images are segmented to obtain a set of physical images, and then the set of physical images is constructed. Model processing is performed to obtain the physical three-dimensional model, the first three-dimensional scene model is generated according to the physical three-dimensional model and the environmental three-dimensional model, and finally the first three-dimensional scene model is sent to the client. Therefore, the embodiment of the present application processes the environment image collection sent by the client. Carry out modeling processing to meet customer needs and achieve remote customization effects.
  • FIG. 7 is a three-dimensional model transmission method provided by another embodiment of the present application.
  • the three-dimensional model transmission method may include but is not limited to step S710 and step S720.
  • Step S710 Establish an augmented reality video call with the design end.
  • the design end sends a video conference invitation to the client.
  • the design end can establish an augmented reality video call with the client only after confirmation by the server-side queuing system and the server-side authentication system; or , the client sends a video conference invitation to the design end.
  • the client can establish an augmented reality video call with the design end after being confirmed by the server-side queuing system and the server-side authentication system. There are no specific restrictions here. .
  • Step S720 Receive the first three-dimensional scene model sent by the server.
  • the first three-dimensional scene model is obtained by the server based on the physical three-dimensional model and the environmental three-dimensional model.
  • the physical three-dimensional model is obtained by the server through modeling processing of the physical image collection
  • the environmental three-dimensional model is obtained by the server.
  • the environmental image collection is obtained through modeling processing. Both the physical image collection and the environmental image collection are obtained by segmenting multiple video images on the server side.
  • the video images are extracted by the design end by extracting the video stream of the augmented reality video call. obtained through frame processing.
  • edge contour scanning may be used to segment multiple video images to obtain a set of physical images and a set of environmental images; or , the deep learning algorithm can be used to identify and process the physical object images and environmental images in multiple video images, and the segmented areas corresponding to the recognized physical images and the environmental images corresponding to the multiple video images are determined respectively according to the composition rules.
  • segmentation area and then obtain the physical image set and the environmental image set based on the segmentation area corresponding to the physical image and the segmentation area corresponding to the environmental image, wherein the composition rule may include placing the physical image or the environmental image in the segmentation area
  • the rules for setting the location and area occupied are not specifically limited here.
  • the design end can intercept a clear frame of image from the video, or it can intercept multiple image sets, that is, the design end can intercept the video stream of the augmented reality video call. Perform frame extraction processing to obtain multiple video images.
  • audio and video switching can be supported.
  • the client can switch to the playback interface of application A and display the augmented reality video.
  • the call is set as a background program so that the augmented reality video call runs in the background.
  • Application A can be audio or video.
  • the client when the client is having an augmented reality video call with the designer, the client can There are no specific restrictions on switching a video call to a voice call.
  • the video stream of the augmented reality video call adopts YUV encoding and H.264 video protocol, where YUV is a color encoding format, "Y” represents brightness, that is, grayscale value; "U” and “ V” represents chroma. Both "U” and “V” can describe image color and saturation, and can be used to specify the color of pixels.
  • the H.264 video protocol is a digital video compression encoding standard.
  • the video stream of the augmented reality video call can be a video stream on the client side, a video stream on the design side, or a video stream on the client side and a video stream on the user side. It is obtained based on the actual situation and there are no specific restrictions here.
  • the design end can extract frames from the video stream on the design end to obtain multiple video images; or, after the design end and the client establish an augmented reality video call, , the design end can perform frame extraction processing on the video stream on the client side to obtain multiple video images; or, after the design end and the client establish an augmented reality video call, the design end can process the video stream on the design end side and the client side respectively.
  • the video stream on the client side undergoes frame extraction processing to obtain multiple video images.
  • the client by adopting the three-dimensional model transmission method including the above steps S710 to S720, first the client establishes an augmented reality video call with the design terminal, and then the client receives the first three-dimensional scene model sent by the server, wherein the third A three-dimensional scene model is obtained by the server side based on the physical three-dimensional model and the environmental three-dimensional model.
  • the physical three-dimensional model is obtained by the server side by modeling a collection of physical objects.
  • the environmental three-dimensional model is obtained by the server side by modeling a collection of environmental images.
  • the physical image collection and the environment image collection are obtained by segmenting multiple video images on the server side, and the video images are obtained by the design end by extracting frames from the video stream of the augmented reality video call, that is, Say, in the scenario of an augmented reality video call established between the design end and the client, the design end performs frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images, and the server end processes the multiple video images. , finally obtain the first three-dimensional scene model, and finally the client receives the first three-dimensional scene model sent by the server. Therefore, the embodiment of the present application can realize the transmission of the three-dimensional model in the augmented reality scene.
  • the three-dimensional model transmission method may include but is not limited to step S810 and step S820.
  • Step S810 Perform virtual annotation processing on the first three-dimensional scene model to obtain annotation information.
  • the annotation information may include coordinate information, text information, annotated models, etc., and is not specifically limited here.
  • the server may send the first three-dimensional scene model to the design end, and the design end may store and process the first three-dimensional scene model.
  • the client can send the annotation information to the design end in real time. For example, the client can switch the perspective of the first three-dimensional scene model.
  • the client can switch the first three-dimensional scene model to the bedroom area and perform virtual annotation processing on the bedroom area, or it can switch the first three-dimensional scene model to the living room area and Perform virtual annotation processing on the living room area, switch the first three-dimensional scene model to the bathroom area, perform virtual annotation processing on the bathroom area, and finally send the annotation information of all areas to the design end (or server end).
  • the design end can receive the annotation information of all areas in real time, and modify the design of the first three-dimensional scene model stored in the design end based on the annotation information to realize remote modification of the three-dimensional model and improve the user experience. This will not be detailed here. limit.
  • Step S820 Send the annotation information to the server.
  • the client can also send the standard information to the design end, and the design end modifies the design of the first three-dimensional scene model stored in the design end based on the standard information. Since each module in the first three-dimensional scene model Different designers may be responsible for it, that is, multiple designers may make concurrent modifications to the first three-dimensional scene model. Therefore, concurrency conflicts may occur. In order to avoid the problem of concurrency conflicts, multiple designers in the design end can modify the source data files corresponding to the modules based on their own design copies. Finally, the design end submits all modified source data files to the server end. The server collects the source data files and then delivers them to the user in a unified manner. Therefore, in this embodiment of the present application, the first three-dimensional scene model is directly modified based on the annotation information, thereby reducing communication costs and achieving rapid delivery.
  • the annotation information is saved in a relevant file, packaged into a data packet and sent to the server. After receiving the data packet, the server parses the data packet to obtain the annotation information, which will not be detailed here. limit.
  • the client can perform virtual annotation processing on the first three-dimensional scene model, obtain annotation information, and send the annotation information to the server, so that the The server determines the annotation area in the first three-dimensional scene model based on the annotation information, nests the annotation information on the annotation area, and obtains the second three-dimensional scene model.
  • the first three-dimensional scene model and the second three-dimensional scene model are superimposed to obtain
  • the third three-dimensional scene model means that the client can feed back the annotation information to the server, and the server can modify the first three-dimensional scene model based on the annotation information, thereby reducing communication costs, while also helping to meet user needs and improve user experience.
  • the user marks building B in the AR device, and sends the building B to the server through the AR device, and the server parses the collection of physical images. , obtain the graphic information, obtain the physical image element information from the physical model library according to the graphic information, then use the physical image element information to construct the physical image base map, perform modeling processing on the physical image base map, obtain the physical three-dimensional model, and finally convert the physical object
  • the three-dimensional model is sent to the client, and the user also sends the building B to the design end through the AR device.
  • the design end obtains the historical data related to building B from the data storage module based on the building B and saves the historical data. Send it to the AR device.
  • the design end can send the historical data to the AR device through voice, video or text.
  • the user can also control the physical objects in the AR device through voice control or peripheral input.
  • the three-dimensional model is enlarged or reduced, and the viewing angle of the physical three-dimensional model in the AR device is switched to view the model details.
  • the AR device is the client and is not specifically limited here.
  • the design end and the client establish an augmented reality video call, and then the design end performs frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images, and then processes the multiple videos.
  • the image is segmented to obtain a set of physical images and a set of environmental images, and then the set of physical images and the set of environmental images are analyzed and processed respectively.
  • the set of physical images can be analyzed and processed to obtain graphic information.
  • the environmental image element information is obtained from the environmental model library according to the graphic information, use the physical image element information to construct the physical image base map, perform modeling processing on the physical image base map, and obtain the physical three-dimensional model; for the analysis and processing of the environmental image collection, you can
  • the environmental image collection is analyzed and processed to obtain label information.
  • the environmental image element information is obtained from the environmental model library according to the label information.
  • the environmental image element information is used to model the environmental image collection to obtain an environmental three-dimensional model. Then, the environmental three-dimensional model is obtained based on the physical three-dimensional model and the environmental three-dimensional model.
  • the model generates a first three-dimensional scene model, and sends the first three-dimensional scene model to the client, and the client performs three-dimensional rendering processing on the first three-dimensional scene model; or, the server performs three-dimensional rendering processing on the first three-dimensional scene model to obtain the rendered third A three-dimensional scene model, and then the rendered first three-dimensional scene model is sent to the client.
  • the client performs three-dimensional rendering processing on the rendered first three-dimensional scene model again to obtain the re-rendered first three-dimensional scene model.
  • the client performs virtual annotation processing on the rendered first three-dimensional scene model or the re-rendered first three-dimensional scene model to obtain annotation information, and the client applies the annotation information to the rendered first three-dimensional scene model or the re-rendered first three-dimensional scene model.
  • the first three-dimensional scene model is modified concurrently, that is, multiple designers in the design end modify the source data files corresponding to the module based on their own design copies, and finally the design end submits all modified source data files to the server
  • the server side sends the source data files to the client in a unified manner to achieve remote delivery.
  • an embodiment of the present application also provides a three-dimensional model transmission device.
  • the three-dimensional model transmission device includes a design terminal 100, a client 300, and a server terminal 200.
  • the design end 100 includes a first display module 101, a first three-dimensional rendering module 102, a first audio and video processing module 103, a data storage module 104 and an agent integration interface 105.
  • the first display module 101 may be used to display the rendered first three-dimensional scene model or the third three-dimensional scene model, or to display the re-rendered first three-dimensional scene model or the third three-dimensional scene model, wherein the display
  • the module includes a computer display screen or an AR screen, etc.
  • the first three-dimensional rendering module 102 can be used to perform three-dimensional rendering processing on the first three-dimensional scene model or the third three-dimensional scene model to obtain the rendered first three-dimensional scene model or the rendered third three-dimensional scene model.
  • the three-dimensional scene model can also be used to perform three-dimensional rendering processing on the rendered first three-dimensional scene model or the rendered third three-dimensional scene model to obtain the first three-dimensional scene model after re-rendering or the third three-dimensional scene model after re-rendering.
  • the first audio and video processing module 103 can be used to receive the video stream of the augmented reality video call sent to the agent integration interface 105, and perform frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images, It can also be used to receive the first three-dimensional scene model or the third three-dimensional scene model from the agent integration interface 105, or to receive the rendered first three-dimensional scene model or the rendered third three-dimensional scene model from the agent integration interface 105; data
  • the storage module 104 may be used to store the video stream of the augmented reality video call from the agent integration interface 105, annotation information, model-related data, the first three-dimensional scene model, the third three-dimensional scene model, the rendered first three-dimensional scene model and the rendering.
  • the third three-dimensional scene model, etc.; the agent integration interface 105 can be used to obtain multiple video images in the first audio and video processing module 103, and send the multiple video images to the server 200, and can also obtain the server 200
  • the first audio and video processing module 103 can support audio and video encoding formats; can call the data storage module 104; can sort video conference invitations according to priority, and can select video responses or audio responses for video conference invitations. ;Support multi-stream media negotiation mode simultaneously; The first three-dimensional scene model or the third three-dimensional scene model can be transmitted to the first three-dimensional rendering module 102.
  • the agent integration interface 105 can be used to receive an access request from the message processing module 203 of the server 200, parse the access request, and also obtain relevant information from the first audio and video processing module 103 for packaging. Obtain the data packet and send the data packet to the server 200, which is not specifically limited here.
  • the server 200 includes a file storage 201, a message processing module 203, a message cache module 204, a data flow module 205, an image analysis module 209, a model adaptation module 210, a three-dimensional model constructor 211, a model storage module 208 and a second three-dimensional rendering module. 206.
  • the file storage 201 can be used to store multiple video images from the agent integration interface 105; the message processing module 203 can be used to authenticate and confirm the video conference invitation; the message caching module 204 can be used to receive users of the message processing module 203 Information and data such as agent addresses; the data flow module 205 can be used to transmit the video stream of the augmented reality video call from the agent integration interface 105 of the design end 100, and can be used to send video invitations to join the meeting and exit the augmented reality video call; image The parsing module 209 can be used to parse the physical image set to obtain graphic information, and to parse the environmental image set to obtain label information; the model adaptation module 210 can be used to construct the physical image base map using the physical image element information, and The physical image base map is subjected to modeling processing to obtain a three-dimensional model of the physical object; the model storage module 208 can be used to obtain environmental image element information from the environmental model library according to the label information, and can also be used to obtain the physical object from the physical model
  • Image element information; the three-dimensional model constructor 211 can be used to perform modeling processing on the environment image set using the environment image element information to obtain an environment three-dimensional model; the second three-dimensional rendering module 206 can be used to convert the first three-dimensional model constructor 211 from the three-dimensional model constructor 211.
  • the three-dimensional scene model or the third three-dimensional scene model is subjected to three-dimensional rendering processing to obtain the rendered first three-dimensional scene model or the third three-dimensional scene model.
  • the file storage 201 can also receive a data package from the agent integration interface 105 through the upload and download module 202, and store it in a designated directory, where the data package includes a three-dimensional model information file.
  • the server 200 also includes an upload and download module 202.
  • the upload and download module can be used to detect the number of directory levels according to the directory of files stored in the file storage 201, detect whether the directory of the server 200 exists, and detect the file. Whether the name or format is correct, and the token is legal. After confirming that it is correct, generate a URL (Uniform Resource Locator, Uniform Resource Locator) address, and send the URL address to the design end 100 and the client 300.
  • the upload and download module can be used to detect the number of directory levels according to the directory of files stored in the file storage 201, detect whether the directory of the server 200 exists, and detect the file. Whether the name or format is correct, and the token is legal. After confirming that it is correct, generate a URL (Uniform Resource Locator, Uniform Resource Locator) address, and send the URL address to the design end 100 and the client 300.
  • URL Uniform Resource Locator
  • the message processing module 203 can forward the account login request; can distribute the query operation initiated by the client 300 to other modules, and send messages to other modules; at the same time, when there are too many request messages to be processed, one can Data such as user information and agent address are stored in the message cache module 204.
  • the image analysis module 209 can also apply image processing algorithms and computer vision algorithms, and the image analysis model can also be used for image recognition and image segmentation.
  • a tape from the design end 100 After being processed by the image parsing module 209, the object marked with a red frame parses out the graphic information of the object, such as the object's sub-model and the marked coordinate information, and then notifies the client 300 through the message processing module 203 and the message caching module 204. Capture the object in the video stream.
  • the image analysis module 209 can obtain multiple video images from the file storage 201 and perform segmentation processing on the multiple video images to obtain a set of physical images and a set of environmental images.
  • the model storage module 208 includes a preset three-dimensional model library, and the preset three-dimensional model library includes a physical model library and an environment model library.
  • the first audio and video processing module 103 of the design end 100 can obtain the video stream of the augmented reality video call from the agent integration interface 105 or from the data storage.
  • the first audio and video processing module 103 performs frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images, and then the first audio and video processing module 103 processes the multiple video images.
  • the video image is sent to the server 200 through the agent integration interface 105.
  • the file storage 201 can obtain multiple video images of the agent integration interface 105 of the design terminal 100 through the upload and download module 202, and can also obtain multiple video images of the client integration interface 302 of the client 300 through the upload and download module 202.
  • Video images are not specifically limited here.
  • the file storage 201 stores multiple video images from the agent integration interface 105 of the design end 100, or stores multiple video images from the client integration interface 302 of the client 300.
  • the image parsing module 209 can extract data from the file. Acquire multiple video images in the memory 201, segment the multiple video images to obtain a set of physical images and a set of environmental images, perform analysis on the set of physical images to obtain graphic information, and perform analysis on the set of environmental images to obtain labels. information, and sends the graphic information and the label information to the model adaptation module 210.
  • the model adaptation module 210 uses the graphic information to obtain the physical image element information from the physical model library in the model storage module 208, and uses the physical image element
  • the information is used to construct a base map of the physical image, and modeling is performed on the base map of the physical object to obtain a three-dimensional model of the physical object, and the model adaptation module 210 uses the label information to obtain the environmental image element information from the environment model library in the model storage module 208, using The environmental image element information is used to model the environmental image collection to obtain an environmental three-dimensional model.
  • the model adaptation module 210 sends the physical three-dimensional model and the environmental three-dimensional model to the three-dimensional model constructor 211.
  • the three-dimensional model constructor 211 uses the physical three-dimensional model and the environment.
  • the three-dimensional model generates a first three-dimensional scene model, and sends the first three-dimensional scene model to the second three-dimensional rendering module 206.
  • the second three-dimensional rendering module 206 performs three-dimensional rendering processing on the first three-dimensional scene model to obtain the rendered first three-dimensional scene model.
  • scene model and then the second three-dimensional rendering module 206 sends the rendered first three-dimensional scene model to the video stream module, and sends it to the agent integration interface 105 of the design end 100 through the video stream module; or, the three-dimensional model constructor 211 uploads
  • the download module 202 sends the first three-dimensional scene model to the file storage 201, and the file storage 201 stores the first three-dimensional scene model.
  • the first three-dimensional scene module can also be sent to the agent integration interface 105, or the first three-dimensional scene module can be sent to the agent integration interface 105.
  • the scene module is sent to the client 300 agent interface.
  • the agent integration interface 105 of the design end 100 can obtain the rendered first three-dimensional scene module from the video stream module and send it to the data storage module 104.
  • the first audio and video processing module 103 can obtain the rendered first three-dimensional scene module from the data storage module 104. Obtaining the rendered first three-dimensional scene module, the first audio and video processing module 103 directly sends the rendered first three-dimensional scene module to the first display module 101, and the first display module 101 displays the rendered first three-dimensional scene module.
  • the model is displayed and no specific restrictions are made here.
  • the agent integration interface 105 of the design end 100 can obtain the rendered first three-dimensional scene module from the data storage module 104, and the first audio and video processing module 103 can obtain the rendered first three-dimensional scene module from the data storage module 104.
  • the three-dimensional scene module is sent to the first three-dimensional rendering module 102.
  • the first three-dimensional rendering module 102 performs three-dimensional rendering processing on the rendered first three-dimensional scene model again.
  • the first display module 101 receives another three-dimensional rendering process from the first three-dimensional rendering module 102.
  • the rendered first three-dimensional scene model displays the re-rendered first three-dimensional scene model, which is not specifically limited here.
  • the agent integration interface 105 of the design end 100 can obtain the first three-dimensional scene module from the file storage 201 of the server 200 through the upload and download module 202 of the server 200, and send it to the first audio and video processing module 103.
  • the first audio and video processing module 103 sends the first three-dimensional scene module to the first three-dimensional rendering module 102, and performs three-dimensional rendering processing on the first three-dimensional scene model through the first three-dimensional rendering module 102.
  • the first display module 101 receives from The rendered first three-dimensional scene model of the first three-dimensional rendering module 102 displays the rendered first three-dimensional scene model, which is not specifically limited here.
  • the agent integration interface 105 of the design end 100 can obtain the first three-dimensional scene module from the file storage 201 of the server 200 through the upload and download module 202 of the server 200, and send it to the data storage module 104.
  • the first The audio and video processing module 103 obtains the first three-dimensional scene module from the data storage module 104, sends the first three-dimensional scene module to the first three-dimensional rendering module 102, and performs three-dimensional rendering on the first three-dimensional scene model through the first three-dimensional rendering module 102.
  • the first display module 101 receives the rendered first three-dimensional scene model from the first three-dimensional rendering module 102, and displays the rendered first three-dimensional scene model, which is not specifically limited here.
  • the client 300 can click the cloud rendering button in the 3D program of the local terminal through Web software or directly access the resources through high-speed Internet access.
  • the instructions are issued from the user terminal, and the server executes the corresponding rendering tasks according to the instructions. , and the rendering result screen is sent back to the user terminal for display.
  • Providing remote rendering capabilities to terminal devices can make up for the terminal's shortcomings in rendering capabilities.
  • the server 200 also includes an information database 207, where the information database 207 is a storage database for important information data of the server 200.
  • the message cache module 204 can store some important user information, agent address and other data into the information database 207, and the information database 207 can be used to resume interrupted services.
  • the client 300 includes a client integrated interface 302, a camera acquisition module 303, a second display module 306, a third three-dimensional rendering module 305 and a second audio and video processing module 304.
  • the client integration interface 302 can be used to obtain multiple video images in the second audio and video processing module 304, and send the multiple video images to the server 200;
  • the second display module 306 can be used to display the rendered first A three-dimensional scene model or a third three-dimensional scene model, in which the display module includes a computer display screen or an AR screen.
  • the third three-dimensional rendering module 305 is used to perform three-dimensional rendering processing on the first three-dimensional scene model or the third three-dimensional scene model from the server 200 to obtain the rendered first three-dimensional scene model or the third three-dimensional scene model; the second sound
  • the video processing module 304 can be used to receive the video stream of the augmented reality video call sent to the client integrated interface 302, and perform frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images; the camera acquisition module 303 is used to Capture the video stream of an augmented reality video call.
  • the client 300 also includes a local storage module 301.
  • the local storage module 301 is used to upload or download files from the file storage 201 of the server 200, according to the URL address sent by the upload and download module 202, to the server. 200 files are saved in the local address or uploaded to the server 200.
  • the format of the files can also be converted, such as converting pictures into binary files.
  • the client integration interface 302 can be used to receive an access request from the message processing module 203 of the server 200, parse the access request, and also obtain relevant information in the first audio and video processing module 103 for packaging. Obtain the data packet and send the data packet to the server 200, which is not specifically limited here.
  • the local storage module 301 of the client 300 can download the first three-dimensional scene module from the file storage 201 of the server 200 through the upload and download module 202 of the server 200.
  • the camera collection module 303 collects the video stream of the augmented reality video call, and sends the video stream to the second audio and video processing module 304.
  • the second audio and video processing module 304 processes the video stream of the augmented reality video call.
  • Frame extraction processing is performed to obtain multiple video images, and the multiple video images are sent to the client integration interface 302.
  • the client integration interface 302 sends the multiple video images to the server 200.
  • the client integration interface 302 sends the multiple video images to the server 200.
  • the multiple video images are stored in the local storage module 301, and the multiple video images are sent to the file storage 201 of the server 200 through the upload and download module 202 of the server 200 through the local storage module 301.
  • the client integration interface 302 of the client 300 can obtain the rendered first three-dimensional scene module from the video stream module and send it to the second audio and video processing module 304.
  • the second audio and video processing module 304 will Rendered
  • the first three-dimensional scene module is directly sent to the second display module 306, and the second display module 306 displays the rendered first three-dimensional scene model, which is not specifically limited here.
  • the client integration interface 302 of the client 300 can obtain the rendered first three-dimensional scene module from the video stream module and send it to the second audio and video processing module 304.
  • the second audio and video processing module 304 will The rendered first three-dimensional scene module is sent to the third three-dimensional rendering module 305, and the third three-dimensional rendering module 305 performs three-dimensional rendering processing on the rendered first three-dimensional scene model again.
  • the second display module 306 receives data from the third three-dimensional rendering module 305.
  • the re-rendered first three-dimensional scene model of the three-dimensional rendering module 305 displays the re-rendered first three-dimensional scene model, which is not specifically limited here.
  • the client integration interface 302 of the client 300 can obtain the first three-dimensional scene module from the local storage module 301 and send it to the second audio and video processing module 304, and the second audio and video processing module 304 will process the first three-dimensional scene module.
  • the three-dimensional scene module is sent to the third three-dimensional rendering module 305, and the third three-dimensional rendering module 305 performs three-dimensional rendering processing on the first three-dimensional scene model.
  • the second display module 306 receives the rendered third image from the third three-dimensional rendering module 305.
  • a three-dimensional scene model is used to display the rendered first three-dimensional scene model, which is not specifically limited here.
  • the first step is to perform relevant processing.
  • the user's permission or consent is obtained, and the collection, use and processing of this data will comply with the relevant laws, regulations and standards of the relevant countries and regions.
  • the embodiment of this application needs to obtain the user's sensitive personal information, it will obtain the user's separate permission or separate consent through a pop-up window or jump to a confirmation page. After clearly obtaining the user's separate permission or separate consent, it will then Obtain necessary user-related data for normal operation of the embodiment of the present application.
  • the three-dimensional model transmission device 400 includes a memory 402, a processor 401, and is stored in the memory 402 and can be run on the processor 401. computer program.
  • the processor 401 and the memory 402 may be connected through a bus or other means.
  • the memory 402 can be used to store non-transitory software programs and non-transitory computer executable programs.
  • memory 402 may include high-speed random access memory and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid-state storage device.
  • the memory 402 may include memory located remotely relative to the processor 401, and these remote memories may be connected to the processor 401 through a network. Examples of the above-mentioned networks include but are not limited to the Internet, intranets, local area networks, mobile communication networks and combinations thereof.
  • the non-transient software programs and instructions required to implement the three-dimensional model transmission method in the above embodiment are stored in the memory 402.
  • the three-dimensional model transmission method in the above embodiment is executed, for example, the above-described The method steps S110 to S160 in Figure 1, the method steps S210 to S230 in Figure 2, the method steps S310 to S350 in Figure 3, the method steps S410 to S440 in Figure 4, the method steps S510 to S530 in Figure 5, Method steps S610 to S620 in FIG. 6 , method steps S710 to S720 in FIG. 7 , and method steps S810 to S820 in FIG. 8 .
  • the device embodiments described above are only illustrative, and the units described as separate components may or may not be physically separate, that is, they may be located in one place, or they may be distributed to multiple network units. Some or all of the modules can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
  • an embodiment of the present application also provides a computer-readable storage medium that stores computer-executable instructions, and the computer-executable instructions are executed by a processor or controller, for example, by the above-mentioned Execution by a processor in the device embodiment can cause the above-mentioned processor to execute the three-dimensional model transmission method in the above embodiment, and execute the above-described method steps S110 to S160 in Figure 1, method steps S210 to S230 in Figure 2, The method steps S310 to S350 in Figure 3, the method steps S410 to S440 in Figure 4, the method steps S510 to S530 in Figure 5, the method steps S610 to S620 in Figure 6, the method steps S710 to S720 in Figure 7 and Method steps S810 to S820 in Figure 8 .
  • an embodiment of the present application also provides a computer program product, including a computer program or computer instructions Let the computer program or computer instructions be stored in a computer-readable storage medium, the processor of the computer device reads the computer program or computer instructions from the computer-readable storage medium, and the processor executes the computer program or computer instructions, causing the computer device to perform the above implementation
  • the three-dimensional model transmission method in the example for example, executes the above-described method steps S110 to S160 in Figure 1, method steps S210 to S230 in Figure 2, method steps S310 to S350 in Figure 3, and method steps in Figure 4.
  • the embodiments of this application include: first, the server side obtains multiple video images from the design side.
  • the video images are obtained by the design side by performing frame extraction processing on the video stream of the augmented reality video call.
  • the augmented reality video call is established by the design side and the client. , then perform segmentation processing on multiple video images to obtain a collection of physical images and a collection of environmental images, then perform modeling processing on the collection of physical images to obtain a three-dimensional model of the physical object, and perform modeling processing on the collection of environmental images to obtain a three-dimensional model of the environment,
  • the first 3D scene model is generated based on the physical 3D model and the environment 3D model, and finally the first 3D scene model is sent to the client.
  • the design end In the scenario of the augmented reality video call established between the design end and the client, the design end The video stream of the augmented reality video call is subjected to frame extraction processing to obtain multiple video images.
  • the server processes the multiple video images to finally obtain a first three-dimensional scene model, and sends the first three-dimensional scene model to the client. , Therefore, the embodiments of the present application can realize the transmission of three-dimensional models in augmented reality scenarios.
  • Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, Digital Versatile Disk (DVD) or other optical disk storage, magnetic cassettes, tapes, disk storage or other magnetic storage devices, or may Any other medium used to store the desired information and that can be accessed by a computer.
  • communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism, and may include any information delivery media .

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Graphics (AREA)
  • Software Systems (AREA)
  • Geometry (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Hardware Design (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Processing Or Creating Images (AREA)

Abstract

Provided in the present application are a three-dimensional model transmission method and apparatus, and a storage medium and a program product. The method comprises: acquiring a plurality of video images, wherein the video images are obtained by a design end by means of performing frame extraction processing on a video stream of an augmented reality video call which is established with a client by the design end (S110); performing segmentation processing on the plurality of video images to obtain a real object image set and an environment image set (S120); respectively performing modeling processing on the real object image set and the environment image set, so as to obtain a real object three-dimensional model and an environment three-dimensional model (S130-S140); generating a first three-dimensional scene model according to the real object three-dimensional model and the environment three-dimensional model (S150); and sending the first three-dimensional scene model to the client (S160).

Description

三维模型传输方法及其装置、存储介质、程序产品Three-dimensional model transmission method and device, storage medium, and program product
相关申请的交叉引用Cross-references to related applications
本申请基于申请号为202210640757.1、申请日为2022年06月08日的中国专利申请提出,并要求该中国专利申请的优先权,该中国专利申请的全部内容在此引入本申请作为参考。This application is filed based on a Chinese patent application with application number 202210640757.1 and a filing date of June 8, 2022, and claims the priority of the Chinese patent application. The entire content of the Chinese patent application is hereby incorporated by reference into this application.
技术领域Technical field
本申请实施例涉及但不限于通信技术领域,尤其涉及一种三维模型传输方法及其装置、存储介质、程序产品。Embodiments of the present application relate to but are not limited to the field of communication technology, and in particular, to a three-dimensional model transmission method and its device, storage medium, and program product.
背景技术Background technique
相关技术中,在AR(Augmented Reality,增强现实)场景中,客户端与设计端之间只能传输视频、文本和图片等文件,而不能传输三维(3D)模型。因此,如何在增强现实场景下实现三维模型的传输,是亟待解决的一个问题。In related technologies, in AR (Augmented Reality, augmented reality) scenarios, only files such as videos, texts, and pictures can be transmitted between the client and the design end, but not three-dimensional (3D) models. Therefore, how to realize the transmission of 3D models in augmented reality scenes is an issue that needs to be solved urgently.
发明内容Contents of the invention
本申请实施例提供了一种三维模型传输方法及其装置、存储介质、程序产品。Embodiments of the present application provide a three-dimensional model transmission method and its device, storage media, and program products.
第一方面,本申请实施例提供了一种三维模型传输方法,包括:获取来自设计端的多个视频图像,所述视频图像由所述设计端通过对增强现实视频通话的视频流进行抽帧处理而获得,所述增强现实视频通话由所述设计端与客户端建立;对多个所述视频图像进行分割处理,得到实物图像集合和环境图像集合;对所述实物图像集合进行建模处理,得到实物三维模型;对所述环境图像集合进行建模处理,得到环境三维模型;根据所述实物三维模型和所述环境三维模型生成第一三维场景模型;将所述第一三维场景模型发送至所述客户端。In the first aspect, embodiments of the present application provide a three-dimensional model transmission method, including: acquiring multiple video images from the design end, and the video images are processed by the design end by performing frame extraction on the video stream of the augmented reality video call. To obtain, the augmented reality video call is established by the design terminal and the client; perform segmentation processing on a plurality of the video images to obtain a set of physical images and a set of environmental images; perform modeling processing on the set of physical images, Obtain the physical three-dimensional model; perform modeling processing on the environmental image collection to obtain the environmental three-dimensional model; generate a first three-dimensional scene model according to the physical three-dimensional model and the environmental three-dimensional model; send the first three-dimensional scene model to the client.
第二方面,本申请实施例提供了一种三维模型传输方法,包括:与设计端建立增强现实视频通话;接收服务器端发送的第一三维场景模型,所述第一三维场景模型由所述服务器端根据实物三维模型和环境三维模型而获得,所述实物三维模型由所述服务器端通过对实物图像集合进行建模处理而获得,所述环境三维模型由所述服务器端通过对环境图像集合进行建模处理而获得,所述实物图像集合和所述环境图像集合均由所述服务器端通过对多个视频图像进行分割处理而获得,所述视频图像由所述设计端通过对增强现实视频通话的视频流进行抽帧处理而获得。In the second aspect, embodiments of the present application provide a three-dimensional model transmission method, which includes: establishing an augmented reality video call with the design end; receiving a first three-dimensional scene model sent by the server, and the first three-dimensional scene model is sent by the server The terminal is obtained based on the physical three-dimensional model and the environmental three-dimensional model. The physical three-dimensional model is obtained by the server side by modeling the physical object image collection. The environmental three-dimensional model is obtained by the server side by modeling the environmental image collection. The physical image set and the environment image set are obtained by the server side by segmenting multiple video images, and the video images are obtained by the design end through an augmented reality video call. The video stream is obtained by performing frame extraction processing.
第三方面,本申请实施例还提供了一种三维模型传输装置,包括:存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现如上所述的三维模型传输方法。In a third aspect, embodiments of the present application also provide a three-dimensional model transmission device, including: a memory, a processor, and a computer program stored in the memory and executable on the processor. When the processor executes the computer program Implement the three-dimensional model transfer method as described above.
第四方面,本申请实施例还提供了一种计算机可读存储介质,存储有计算机可执行指令,所述计算机可执行指令用于执行如上所述的三维模型传输方法。In a fourth aspect, embodiments of the present application also provide a computer-readable storage medium that stores computer-executable instructions, and the computer-executable instructions are used to execute the three-dimensional model transmission method as described above.
第五方面,本申请实施例还提供了一种计算机程序产品,包括计算机程序或计算机指令,所述计算机程序或所述计算机指令存储在计算机可读存储介质中,计算机设备的处理器从所述计算机可读存储介质读取所述计算机程序或所述计算机指令,所述处理器执行所述计算机程序或所述计算机指令,使得所述计算机设备执行如上所述的三维模型传输方法。In a fifth aspect, embodiments of the present application further provide a computer program product, which includes a computer program or computer instructions. The computer program or computer instructions are stored in a computer-readable storage medium. The processor of the computer device obtains the information from the computer program or computer instructions. The computer-readable storage medium reads the computer program or the computer instructions, and the processor executes the computer program or the computer instructions, so that the computer device performs the three-dimensional model transmission method as described above.
附图说明Description of the drawings
图1是本申请一个实施例提供的三维模型传输方法的流程图;Figure 1 is a flow chart of a three-dimensional model transmission method provided by an embodiment of the present application;
图2是图1中步骤S150的一种方法的流程图;Figure 2 is a flow chart of a method in step S150 in Figure 1;
图3是本申请另一个实施例提供的三维模型传输方法的流程图; Figure 3 is a flow chart of a three-dimensional model transmission method provided by another embodiment of the present application;
图4是图1中步骤S130的一种方法的流程图;Figure 4 is a flow chart of a method in step S130 in Figure 1;
图5是图1中步骤S140的一种方法的流程图;Figure 5 is a flow chart of a method in step S140 in Figure 1;
图6是本申请另一个实施例提供的三维模型传输方法的流程图;Figure 6 is a flow chart of a three-dimensional model transmission method provided by another embodiment of the present application;
图7是本申请另一个实施例提供的三维模型传输方法的流程图;Figure 7 is a flow chart of a three-dimensional model transmission method provided by another embodiment of the present application;
图8是本申请另一个实施例提供的三维模型传输方法的流程图;Figure 8 is a flow chart of a three-dimensional model transmission method provided by another embodiment of the present application;
图9是本申请一个实施例提供的三维模型传输方法的流程图;Figure 9 is a flow chart of a three-dimensional model transmission method provided by an embodiment of the present application;
图10是本申请一个实施例提供的三维模型传输装置的结构示意图;Figure 10 is a schematic structural diagram of a three-dimensional model transmission device provided by an embodiment of the present application;
图11是本申请另一个实施例提供的三维模型传输装置的结构示意图。Figure 11 is a schematic structural diagram of a three-dimensional model transmission device provided by another embodiment of the present application.
具体实施方式Detailed ways
为了使本申请的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处所描述的实施例仅用以解释本申请,并不用于限定本申请。In order to make the purpose, technical solutions and advantages of the present application more clear, the present application will be further described in detail below with reference to the drawings and embodiments. It should be understood that the embodiments described here are only used to explain the present application and are not used to limit the present application.
需要说明的是,虽然在流程图中示出了逻辑顺序,但是在某些情况下,可以以不同于流程图中的顺序执行所示出或描述的步骤。说明书和权利要求书及上述附图的描述中,多个(或多项)的含义是两个以上,大于、小于、超过等理解为不包括本数,以上、以下、以内等理解为包括本数。如果有描述到“第一”、“第二”等只是用于区分技术特征为目的,而不能理解为指示或暗示相对重要性或者隐含指明所指示的技术特征的数量或者隐含指明所指示的技术特征的先后关系。It should be noted that although a logical sequence is shown in the flowchart, in some cases, the steps shown or described may be performed in an order different from that in the flowchart. In the description of the specification, claims and the above drawings, plural (or multiple) means two or more, greater than, less than, exceeding, etc. are understood to exclude the number, and above, below, within, etc. are understood to include the number. If there are descriptions of "first", "second", etc., they are only used for the purpose of distinguishing technical features and cannot be understood as indicating or implying the relative importance or implicitly indicating the number of technical features indicated or implicitly indicating the indicated technical features. The sequence relationship of technical features.
本申请提供了一种三维模型传输方法及其装置、存储介质、程序产品,首先服务器端获取来自设计端的多个视频图像,视频图像由设计端通过对增强现实视频通话的视频流进行抽帧处理而获得,增强现实视频通话由设计端与客户端建立,接着对多个视频图像进行分割处理,得到实物图像集合和环境图像集合,然后对实物图像集合进行建模处理,得到实物三维模型,并且对环境图像集合进行建模处理,得到环境三维模型,根据实物三维模型和环境三维模型生成第一三维场景模型,最后将第一三维场景模型发送至客户端,即是说,在设计端与客户端建立的增强现实视频通话的场景下,设计端对该增强现实视频通话的视频流进行抽帧处理得到多个视频图像,服务器端通过对该多个视频图像进行处理,最终得到第一三维场景模型,并将该第一三维场景模型发送至客户端,因此,本申请实施例能够在增强现实场景下实现三维模型的传输。This application provides a three-dimensional model transmission method and its device, storage media, and program products. First, the server side obtains multiple video images from the design side. The video images are processed by the design side by extracting frames from the video stream of the augmented reality video call. To obtain, the augmented reality video call is established between the design end and the client, and then multiple video images are segmented to obtain a collection of physical images and a collection of environmental images, and then the collection of physical images is modeled to obtain a three-dimensional model of the physical object, and The environment image collection is modeled and processed to obtain an environment three-dimensional model. The first three-dimensional scene model is generated based on the physical three-dimensional model and the environment three-dimensional model. Finally, the first three-dimensional scene model is sent to the client. That is to say, at the design end and the client In the scenario of an augmented reality video call created on the client, the design end performs frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images. The server end processes the multiple video images to finally obtain the first three-dimensional scene. model, and sends the first three-dimensional scene model to the client. Therefore, the embodiment of the present application can realize the transmission of the three-dimensional model in the augmented reality scene.
下面结合附图,对本申请实施例作进一步阐述。The embodiments of the present application will be further described below with reference to the accompanying drawings.
参照图1,图1是本申请一个实施例提供的三维模型传输方法的流程图,该三维模型传输方法可以包括但不限于步骤S110、步骤S120、步骤S130、步骤S140、步骤S150和步骤S160。Referring to Figure 1, Figure 1 is a flow chart of a three-dimensional model transmission method provided by an embodiment of the present application. The three-dimensional model transmission method may include but is not limited to step S110, step S120, step S130, step S140, step S150 and step S160.
步骤S110:获取来自设计端的多个视频图像。Step S110: Obtain multiple video images from the design end.
在一实施方式中,视频图像由设计端通过对增强现实视频通话的视频流进行抽帧处理而获得,增强现实视频通话由设计端与客户端建立。In one implementation, the video image is obtained by the design end by performing frame extraction processing on the video stream of the augmented reality video call, and the augmented reality video call is established by the design end and the client.
可以理解的是,由于视频的本质是一帧帧连续的图片,所以设计端可以从视频中截取一帧清晰图像,也可以截取多张图像集,即设计端通过对增强现实视频通话的视频流进行抽帧处理,得到多个视频图像。It is understandable that since the essence of video is continuous pictures frame by frame, the design end can intercept a clear frame of image from the video, or it can intercept multiple image sets, that is, the design end can intercept the video stream of the augmented reality video call. Perform frame extraction processing to obtain multiple video images.
在一实施方式中,设计端向客户端发送视频入会邀请,客户端确认接受后,经过服务器端鉴权确认,设计端和客户端才可以成功建立增强现实视频通话;或者,客户端向设计端发送视频入会邀请,设计端确认接受后,经过服务器端的排队系统和服务器端的鉴权系统的确认,设计端和客户端才可以成功建立增强现实视频通话,在此不做具体限制。In one implementation, the design end sends a video conference invitation to the client. After the client confirms acceptance, the design end and the client can successfully establish an augmented reality video call after server-side authentication and confirmation; or, the client sends a video conference invitation to the design end. After sending a video conference invitation and confirming acceptance on the design side, and after confirmation by the server-side queuing system and server-side authentication system, the design side and the client can successfully establish an augmented reality video call. There are no specific restrictions here.
在一实施方式中,在增强现实视频通话的过程中,可以支持音视频切换,比如,客户端在增强现实视频通话的过程中,客户端可以切换到应用A的播放界面,并将增强现实视频通话置为后台程序,使增强现实视频通话在后台运行,其中,应用A可以是音频,也可以是视 频;又如,客户端在与设计端进行增强现实视频通话的过程中,客户端可以将视频通话切换成语音通话,在此不做具体限制。In one embodiment, during the augmented reality video call, audio and video switching can be supported. For example, during the augmented reality video call, the client can switch to the playback interface of application A and display the augmented reality video. Set the call as a background program so that the augmented reality video call runs in the background. Application A can be audio or video. frequency; for another example, when the client is having an augmented reality video call with the designer, the client can switch the video call to a voice call, and there are no specific restrictions here.
在一实施方式中,增强现实视频通话的视频流可以采用YUV编码和H.264视频协议,其中,YUV为一种颜色编码格式,“Y”表示明亮度,即灰阶值;“U”和“V”表示色度,“U”和“V”均可以描述影像色彩及饱和度,可以用于指定像素的颜色,其中,H.264视频协议为一种数字视频压缩编码标准。In one implementation, the video stream of the augmented reality video call can adopt YUV encoding and H.264 video protocol, where YUV is a color encoding format, "Y" represents brightness, that is, grayscale value; "U" and "V" represents chroma, "U" and "V" can both describe image color and saturation, and can be used to specify the color of pixels. Among them, the H.264 video protocol is a digital video compression encoding standard.
在一实施方式中,增强现实视频通话的视频流可以是客户端侧的视频流,也可以是设计端侧的视频流,也可以是客户端侧的视频流和用户端侧的视频流,可以根据实际情况获取,在此不做具体限制。比如,在设计端与客户端建立增强现实视频通话后,设计端可以对设计端侧的视频流进行抽帧处理,得到多个视频图像;或者,在设计端与客户端建立增强现实视频通话后,设计端可以对客户端侧的视频流进行抽帧处理,得到多个视频图像;又或者,在设计端与客户端建立增强现实视频通话后,设计端可以分别对设计端侧的视频流和客户端侧的视频流进行抽帧处理,得到多个视频图像。In one embodiment, the video stream of the augmented reality video call can be a video stream on the client side, a video stream on the design side, or a video stream on the client side and a video stream on the user side. It is obtained based on the actual situation and there are no specific restrictions here. For example, after the design end and the client establish an augmented reality video call, the design end can extract frames from the video stream on the design end to obtain multiple video images; or, after the design end and the client establish an augmented reality video call, , the design end can perform frame extraction processing on the video stream on the client side to obtain multiple video images; or, after the design end and the client establish an augmented reality video call, the design end can process the video stream on the design end side and the client side respectively. The video stream on the client side undergoes frame extraction processing to obtain multiple video images.
步骤S120:对多个视频图像进行分割处理,得到实物图像集合和环境图像集合。Step S120: Perform segmentation processing on multiple video images to obtain a set of physical images and a set of environmental images.
可以理解的是,实物图像集合包括多个实物图像,环境图像集合包括多个环境图像。It can be understood that the physical object image set includes multiple physical object images, and the environmental image set includes multiple environmental images.
在一实施方式中,对多个视频图像进行分割处理,可以有很多实施方式,比如,可以是采用边缘轮廓扫描的方式对多个视频图像进行分割处理,得到实物图像集合和环境图像集合;或者,可以利用深度学习算法对多个视频图像中的实物图像和环境图像进行识别处理,并根据构图规则在该多个视频图像中分别确定与识别出的实物图像对应的分割区域和环境图像对应的分割区域,再根据该实物图像对应的分割区域和环境图像对应的分割区域,得到的实物图像集合和环境图像集合,其中,该构图规则可以包括将实物图像或者环境图像在分割区域中所处的位置、所占的面积进行设定的规则,在此不做具体限制。In one embodiment, there are many implementation methods for segmenting multiple video images. For example, edge contour scanning may be used to segment multiple video images to obtain a set of physical images and a set of environmental images; or , the deep learning algorithm can be used to identify and process the physical object images and environmental images in multiple video images, and the segmented areas corresponding to the recognized physical images and the environmental images corresponding to the multiple video images are determined respectively according to the composition rules. segmentation area, and then obtain the physical image set and the environmental image set based on the segmentation area corresponding to the physical image and the segmentation area corresponding to the environmental image, wherein the composition rule may include dividing the physical image or the environmental image in the segmentation area. There are no specific restrictions on the rules for setting the location and area occupied.
步骤S130:对实物图像集合进行建模处理,得到实物三维模型。Step S130: Perform modeling processing on the collection of real object images to obtain a three-dimensional model of the real object.
步骤S140:对环境图像集合进行建模处理,得到环境三维模型。Step S140: Perform modeling processing on the environment image collection to obtain a three-dimensional environment model.
步骤S150:根据实物三维模型和环境三维模型生成第一三维场景模型。Step S150: Generate a first three-dimensional scene model based on the physical three-dimensional model and the environmental three-dimensional model.
步骤S160:将第一三维场景模型发送至客户端。Step S160: Send the first three-dimensional scene model to the client.
在一实施方式中,服务器端可以将第一三维场景模型发送至设计端,在此不做具体限制。In one implementation, the server can send the first three-dimensional scene model to the design end, which is not specifically limited here.
本实施例中,通过采用包括有上述步骤S110至步骤S160的三维模型传输方法,首先服务器端获取来自设计端的多个视频图像,视频图像由设计端通过对增强现实视频通话的视频流进行抽帧处理而获得,增强现实视频通话由设计端与客户端建立,接着对多个视频图像进行分割处理,得到实物图像集合和环境图像集合,然后对实物图像集合进行建模处理,得到实物三维模型,并且对环境图像集合进行建模处理,得到环境三维模型,根据实物三维模型和环境三维模型生成第一三维场景模型,最后将第一三维场景模型发送至客户端,即是说,在设计端与客户端建立的增强现实视频通话的场景下,设计端对该增强现实视频通话的视频流进行抽帧处理得到多个视频图像,服务器端通过对该多个视频图像进行处理,最终得到第一三维场景模型,并将该第一三维场景模型发送至客户端,因此,本申请实施例能够在增强现实场景下构建三维模型,并实现三维模型的传输。In this embodiment, by adopting the three-dimensional model transmission method including the above-mentioned steps S110 to S160, first, the server side obtains multiple video images from the design side, and the video images are extracted from the video stream of the augmented reality video call by the design side. Obtained through processing, the augmented reality video call is established between the design end and the client, and then multiple video images are segmented to obtain a collection of physical images and a collection of environmental images, and then the collection of physical images is modeled to obtain a three-dimensional model of the physical object. And perform modeling processing on the environment image collection to obtain the environment three-dimensional model, generate the first three-dimensional scene model based on the physical three-dimensional model and the environment three-dimensional model, and finally send the first three-dimensional scene model to the client, that is to say, on the design side and In the scenario of an augmented reality video call established by the client, the design end performs frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images. The server end processes the multiple video images to finally obtain the first three-dimensional scene model, and sends the first three-dimensional scene model to the client. Therefore, the embodiment of the present application can construct a three-dimensional model in an augmented reality scene and realize the transmission of the three-dimensional model.
在一实施方式中,设计端、服务器端和客户端均可以对第一三维场景模型进行三维渲染处理,在此不做具体限制。比如,客户端将服务器端发送的第一三维场景模型进行三维渲染处理,得到渲染后的第一三维场景模型,接着对该渲染后的第一三维场景模型进行显示;或者,当服务器端将第一三维场景模型发送至设计端,设计端对服务器端发送的第一三维场景模型进行三维渲染处理,得到渲染后的第一三维场景模型,接着对该渲染后的第一三维场景模型进行显示;或者,服务器端先对第一三维场景模型进行三维渲染处理,得到渲染后的第一三维场景模型,然后将该渲染后的第一三维场景模型发送至设计端或者客户端。In one implementation, the design end, the server end and the client can all perform three-dimensional rendering processing on the first three-dimensional scene model, which is not specifically limited here. For example, the client performs three-dimensional rendering processing on the first three-dimensional scene model sent by the server to obtain the rendered first three-dimensional scene model, and then displays the rendered first three-dimensional scene model; or, when the server sends the first three-dimensional scene model, the client performs three-dimensional rendering processing on the first three-dimensional scene model sent by the server. A three-dimensional scene model is sent to the design end, and the design end performs three-dimensional rendering processing on the first three-dimensional scene model sent by the server end to obtain the rendered first three-dimensional scene model, and then displays the rendered first three-dimensional scene model; Alternatively, the server first performs three-dimensional rendering processing on the first three-dimensional scene model to obtain the rendered first three-dimensional scene model, and then sends the rendered first three-dimensional scene model to the design end or the client.
在一实施方式中,可以多次对第一三维场景模型进行三维渲染处理,可以根据第一三维场景模型的模型规模确定是否需要多次对第一三维场景模型进行三维渲染处理。举一示例, 当第一三维场景模型模拟的是一个游乐园,那么可以多次对第一三维场景模型进行三维渲染处理,即服务器端先对第一三维场景模型进行三维渲染处理,得到渲染后的第一三维场景模型,然后将该渲染后的第一三维场景模型发送至设计端(或者客户端),设计端(或者客户端)对服务器端发送的该渲染后的第一三维场景模型再次进行三维渲染处理,得到再次渲染后的第一三维场景模型,将该再次渲染后的第一三维场景模型进行显示;或者,当第一三维场景模型模拟的是一个卧室,那么可以只对第一三维场景模型进行一次三维渲染处理,而且可以由设计端、服务器端或者客户端对第一三维场景模型进行三维渲染处理,在此不做具体限制。In one embodiment, the first three-dimensional scene model can be subjected to three-dimensional rendering processing multiple times, and whether it is necessary to perform three-dimensional rendering processing on the first three-dimensional scene model multiple times can be determined according to the model scale of the first three-dimensional scene model. To give an example, When the first three-dimensional scene model simulates an amusement park, the first three-dimensional scene model can be subjected to three-dimensional rendering processing multiple times. That is, the server first performs three-dimensional rendering processing on the first three-dimensional scene model to obtain the rendered first three-dimensional scene model. The scene model is then sent to the design end (or client), and the design end (or client) performs three-dimensional rendering processing again on the rendered first three-dimensional scene model sent by the server. , obtain the re-rendered first three-dimensional scene model, and display the re-rendered first three-dimensional scene model; or, when the first three-dimensional scene model simulates a bedroom, then only the first three-dimensional scene model can be A three-dimensional rendering process is performed once, and the first three-dimensional scene model can be three-dimensionally rendered by the design end, the server end, or the client, and there are no specific restrictions here.
可以理解的是,三维渲染本质上是计算机处理三维模型到二维图像的一个转换过程。渲染涉及到扫描线渲染、光线追踪、光子映射等技术,其可以模拟光与各物质(如材料材质和表面纹理)的相互作用,因此,渲染需要三维插件、三维软件以及硬件的支持。It can be understood that 3D rendering is essentially a conversion process of computer processing of 3D models into 2D images. Rendering involves scan line rendering, ray tracing, photon mapping and other technologies, which can simulate the interaction between light and various substances (such as materials and surface textures). Therefore, rendering requires the support of 3D plug-ins, 3D software and hardware.
在一实施例中,如图2所示,对步骤S150进行进一步的说明,该步骤S150可以包括但不限于有步骤S210、步骤S220和步骤S230。In an embodiment, as shown in FIG. 2 , step S150 is further described. This step S150 may include but is not limited to step S210, step S220 and step S230.
步骤S210:获取预设的三维模型库,三维模型库包括环境模型库。Step S210: Obtain a preset three-dimensional model library, which includes an environment model library.
在一实施方式中,环境模型库中可以包括多个预设环境模型,该多个预设环境模型可以是由模型制作人员使用三维建模工具(比如3D Studio MAX等建模工具,其中,3D Studio MAX是基于电脑操作系统的三维动画渲染和制作软件)对已有模型数据进行修改后,利用修改后的模型数据进行建模处理后得到的,也可以是已有的模型,在此不做具体限制。In one embodiment, the environment model library may include multiple preset environment models. The multiple preset environment models may be created by model makers using three-dimensional modeling tools (such as 3D Studio MAX and other modeling tools, where 3D Studio MAX is a three-dimensional animation rendering and production software based on a computer operating system). It is obtained by modifying the existing model data and using the modified model data for modeling processing. It can also be an existing model, which will not be done here. Specific restrictions.
在一实施方式中,预设的三维模型库可以包括目标三维场景模型、模型相关数据、模型基础部件、模型材质素材等,其中,模型相关数据包括模型规模、模型模拟的实景的位置信息等,在此不做具体限制。例如,在路线规划的场景下,当用户处于山脚下的位置时,用户通过AR设备把位置信息发送给服务器,服务器根据该位置信息从预设的三维模型库中确定该山峰对应的目标三维场景模型,并将该目标三维场景模型发送至设计端,设计端根据该目标三维场景模型对该山峰进行路线规划以及对山路线上的标志性景点、商店、洗手间等进行标记,将该路线规划和标记的相关信息推送给AR客户端,从而提升用户体验,其中AR设备为客户端。In one embodiment, the preset three-dimensional model library may include a target three-dimensional scene model, model-related data, model basic components, model material materials, etc., where the model-related data includes model scale, location information of the real scene simulated by the model, etc., No specific restrictions are made here. For example, in a route planning scenario, when the user is at the foot of a mountain, the user sends location information to the server through the AR device, and the server determines the target 3D scene corresponding to the mountain peak from the preset 3D model library based on the location information. model, and sends the target three-dimensional scene model to the design end. The design end plans the route of the mountain peak based on the target three-dimensional scene model and marks the iconic attractions, shops, restrooms, etc. on the mountain route, and combines the route planning and The relevant information of the mark is pushed to the AR client, thereby improving the user experience, in which the AR device is the client.
步骤S220:根据实物三维模型和环境三维模型从环境模型库中确定目标环境三维模型。Step S220: Determine the target environment three-dimensional model from the environment model library according to the physical three-dimensional model and the environment three-dimensional model.
在一实施例中,可以将实物三维模型和环境三维模型与环境模型库中的预设环境模型进行匹配,确定目标环境三维模型,提高建模效率,同时缩短了建模时间。In one embodiment, the physical three-dimensional model and the environmental three-dimensional model can be matched with the preset environmental model in the environmental model library to determine the target environmental three-dimensional model, thereby improving modeling efficiency and shortening modeling time.
步骤S230:根据实物三维模型和目标环境三维模型生成第一三维场景模型。Step S230: Generate a first three-dimensional scene model based on the physical three-dimensional model and the target environment three-dimensional model.
在本实施例中,通过采用包括上述步骤S210至步骤S230的三维模型传输方法,首先服务器端获取预设的三维模型库,其中,三维模型库包括环境模型库,接着根据实物三维模型和环境三维模型从环境模型库中确定目标环境三维模型,最后根据实物三维模型和目标环境三维模型生成第一三维场景模型,因此,本申请实施例可以通过根据实物三维模型和环境三维模型从环境模型库中确定目标环境三维模型而提高建模效率,同时缩短了建模时间。In this embodiment, by adopting the three-dimensional model transmission method including the above steps S210 to step S230, first the server side obtains a preset three-dimensional model library, where the three-dimensional model library includes an environment model library, and then based on the physical three-dimensional model and the environmental three-dimensional The model determines the target environment three-dimensional model from the environment model library, and finally generates the first three-dimensional scene model based on the physical three-dimensional model and the target environment three-dimensional model. Therefore, the embodiment of the present application can obtain the target environment three-dimensional model from the environment model library based on the physical three-dimensional model and the environment three-dimensional model. Determine the three-dimensional model of the target environment to improve modeling efficiency and shorten modeling time.
在一实施例中,如图3所示,在执行步骤S160之后,该三维模型传输方法还可以包括但不限于步骤S310、步骤S320、步骤S330、步骤S340和步骤S350。In an embodiment, as shown in Figure 3, after executing step S160, the three-dimensional model transmission method may also include but is not limited to step S310, step S320, step S330, step S340 and step S350.
步骤S310:接收客户端发送的标注信息。Step S310: Receive the annotation information sent by the client.
在一实施方式中,标注信息可以包括坐标信息、文字信息、标注过的模型等,而且,该标注信息是由客户端对第一三维场景模型进行虚拟标注处理后得到的。比如,当第一三维场景模型模拟的是一个客厅,且在客厅中坐标为(8,4,0)处放置有沙发,若需要将该沙发换到坐标为(4,4,0)的位置摆放,则坐标信息可以是将坐标为(8,4,0)处的实物三维模型修改到坐标为(4,4,0)的位置。In one embodiment, the annotation information may include coordinate information, text information, annotated models, etc., and the annotation information is obtained by the client performing virtual annotation processing on the first three-dimensional scene model. For example, when the first three-dimensional scene model simulates a living room, and a sofa is placed in the living room at the coordinates (8, 4, 0), if it is necessary to change the sofa to the position at the coordinates (4, 4, 0) If placed, the coordinate information may be to modify the physical three-dimensional model at the coordinates of (8, 4, 0) to the position of the coordinates of (4, 4, 0).
步骤S320:根据标注信息在第一三维场景模型中确定标注区域。Step S320: Determine the labeling area in the first three-dimensional scene model according to the labeling information.
可以理解的是,标注区域可以是修改前的区域,也可以是修改后的区域,也可以是修改前的区域和修改后的区域,比如,标注信息为将坐标为(8,4,0)处的实物三维模型修改到 坐标为(4,4,0)的位置,则标注区域可以是坐标为(8,4,0)的区域,也可以是坐标为(4,4,0)的区域,在此不做具体限制。It can be understood that the labeled area can be the area before modification, the area after modification, or the area before modification and the area after modification. For example, the labeling information is the coordinates (8, 4, 0) The physical 3D model at is modified to If the coordinates are (4, 4, 0), then the marked area can be the area with coordinates (8, 4, 0) or the area with coordinates (4, 4, 0). There are no specific restrictions here. .
步骤S330:在标注区域上嵌套标注信息,得到第二三维场景模型。Step S330: Nest the annotation information on the annotation area to obtain a second three-dimensional scene model.
步骤S340:将第一三维场景模型和第二三维场景模型进行叠加处理,得到第三三维场景模型。Step S340: Overlay the first three-dimensional scene model and the second three-dimensional scene model to obtain a third three-dimensional scene model.
步骤S350:将第三三维场景模型发送至客户端。Step S350: Send the third three-dimensional scene model to the client.
在一实施方式中,服务器端可以将第三三维场景模型发送至设计端,在此不做具体限制。In one implementation, the server side can send the third three-dimensional scene model to the design side, which is not specifically limited here.
在本实施例中,通过采用包括上述步骤S310至步骤S350的三维模型传输方法,首先服务器端接收客户端发送的标注信息,根据标注信息在第三三维场景模型中确定标注区域,然后在标注区域上嵌套标注信息,得到第二三维场景模型,接着将第三三维场景模型和第二三维场景模型进行叠加处理,得到第三三维场景模型,最后将第三三维场景模型发送至客户端,因此,本申请实施例可以接收客户端反馈的标注信息,根据该标注信息对第三三维场景模型进行修改,减少了因专业壁垒导致的沟通成本,实现快速迭代与交付,有利于满足客户需求,同时提高了用户体验。In this embodiment, by adopting the three-dimensional model transmission method including the above-mentioned steps S310 to S350, first the server receives the annotation information sent by the client, determines the annotation area in the third three-dimensional scene model according to the annotation information, and then determines the annotation area in the annotation area. The annotation information is nested on the top to obtain the second three-dimensional scene model. Then the third three-dimensional scene model and the second three-dimensional scene model are superimposed to obtain the third three-dimensional scene model. Finally, the third three-dimensional scene model is sent to the client. Therefore, , the embodiment of this application can receive annotation information fed back by the client, and modify the third three-dimensional scene model based on the annotation information, reducing communication costs caused by professional barriers, achieving rapid iteration and delivery, and conducive to meeting customer needs. At the same time Improved user experience.
在一实施方式中,设计端、服务器端和客户端均可以对第三三维场景模型进行三维渲染处理,在此不做具体限制。比如,客户端将服务器端发送的第三三维场景模型进行三维渲染处理,得到渲染后的第三三维场景模型,接着对该渲染后的第三三维场景模型进行显示;或者,当服务器端将第三三维场景模型发送至设计端,设计端对服务器端发送的第三三维场景模型进行三维渲染处理,得到渲染后的第三三维场景模型,接着对该渲染后的第三三维场景模型进行显示;或者,服务器端先对第三三维场景模型进行三维渲染处理,得到渲染后的第三三维场景模型,然后将该渲染后的第三三维场景模型发送至设计端或者客户端。In one implementation, the design end, the server end and the client can all perform 3D rendering processing on the third 3D scene model, which is not specifically limited here. For example, the client performs three-dimensional rendering processing on the third three-dimensional scene model sent by the server to obtain the rendered third three-dimensional scene model, and then displays the rendered third three-dimensional scene model; or, when the server sends the third three-dimensional scene model, the client performs three-dimensional rendering processing. The three-dimensional scene model is sent to the design end, and the design end performs three-dimensional rendering processing on the third three-dimensional scene model sent by the server to obtain a rendered third three-dimensional scene model, and then displays the rendered third three-dimensional scene model; Alternatively, the server first performs three-dimensional rendering processing on the third three-dimensional scene model to obtain a rendered third three-dimensional scene model, and then sends the rendered third three-dimensional scene model to the design end or the client.
在一实施方式中,可以多次对第三三维场景模型进行三维渲染处理,可以根据第三三维场景模型的模型规模确定是否需要多次对第三三维场景模型进行三维渲染处理。举一示例,当第三三维场景模型模拟的是一个游乐园,那么可以多次对第三三维场景模型进行三维渲染处理,即服务器端先对第三三维场景模型进行三维渲染处理,得到渲染后的第三三维场景模型,然后将该渲染后的第三三维场景模型发送至设计端(或者客户端),设计端(或者客户端)对服务器端发送的该渲染后的第三三维场景模型再次进行三维渲染处理,得到再次渲染后的第三三维场景模型,将该再次渲染后的第三三维场景模型进行显示;或者,当第三三维场景模型模拟的是一个卧室,那么可以只对第三三维场景模型进行一次三维渲染处理,而且可以由设计端、服务器端或者客户端对第三三维场景模型进行三维渲染处理,在此不做具体限制。In one embodiment, the third three-dimensional scene model can be subjected to three-dimensional rendering processing multiple times, and whether it is necessary to perform three-dimensional rendering processing on the third three-dimensional scene model multiple times can be determined according to the model scale of the third three-dimensional scene model. For example, when the third 3D scene model simulates an amusement park, the third 3D scene model can be 3D rendered multiple times. That is, the server first performs 3D rendering on the third 3D scene model and obtains the The third three-dimensional scene model is then sent to the design end (or client), and the design end (or client) responds to the rendered third three-dimensional scene model sent by the server again. Perform three-dimensional rendering processing to obtain a re-rendered third three-dimensional scene model, and display the re-rendered third three-dimensional scene model; or, when the third three-dimensional scene model simulates a bedroom, then only the third three-dimensional scene model can be displayed. The three-dimensional scene model undergoes one three-dimensional rendering process, and the third three-dimensional scene model can be three-dimensionally rendered by the design end, the server end, or the client. There are no specific restrictions here.
在一实施例中,如图4所示,在三维模型库包括实物模型库的情况下,对步骤S130进行进一步的说明,该步骤S130可以包括但不限于有步骤S410、步骤S420、步骤S430和步骤S440。In one embodiment, as shown in Figure 4, when the three-dimensional model library includes a physical model library, step S130 is further described. This step S130 may include but is not limited to step S410, step S420, step S430 and Step S440.
步骤S410:对实物图像集合进行解析处理,得到图形信息。Step S410: Analyze the collection of physical objects to obtain graphic information.
在一实施方式中,图形信息可以包括实物图像的线条轮廓、实物图像位置、实物图像大小以及实物图像的资源地址等,在此不做具体限制。In one embodiment, the graphic information may include the line outline of the physical image, the position of the physical image, the size of the physical image, the resource address of the physical image, etc., which are not specifically limited here.
步骤S420:根据图形信息从实物模型库中获取实物图像元素信息。Step S420: Obtain physical image element information from the physical model library according to the graphic information.
在一实施方式中,实物模型库中可以包括多个预设实物模型,该多个预设实物模型可以是由模型制作人员使用三维建模工具(比如3D Studio MAX等建模工具,其中,3D Studio MAX是基于电脑操作系统的三维动画渲染和制作软件)对已有模型数据进行修改后,利用修改后的模型数据进行建模处理后得到的,也可以是已有的模型,在此不做具体限制。In one embodiment, the physical model library may include multiple preset physical models. The multiple preset physical models may be created by model makers using three-dimensional modeling tools (such as 3D Studio MAX and other modeling tools, where 3D Studio MAX is a three-dimensional animation rendering and production software based on a computer operating system). It is obtained by modifying the existing model data and using the modified model data for modeling processing. It can also be an existing model, which will not be done here. Specific restrictions.
步骤S430:利用实物图像元素信息构建实物图像底图。Step S430: Use the element information of the physical image to construct a base map of the physical image.
在一实施方式中,实物图像底图可以是实物三维模型的基础框架,在此不做具体限制。In one implementation, the physical image base map may be the basic framework of the physical three-dimensional model, which is not specifically limited here.
步骤S440:对实物图像底图进行建模处理,得到实物三维模型。Step S440: Perform modeling processing on the image base map of the physical object to obtain a three-dimensional model of the physical object.
在一实施例中,当该实物图像底图是一个花瓶的轮廓,那么对花瓶进行建模处理,即是 对该花瓶进行材质贴面、调光、纹理、凹陷、凸起处理等,最终得到该花瓶的实物三维模型。In one embodiment, when the physical image base map is the outline of a vase, then the vase is modeled, that is, The vase is subjected to material veneer, light adjustment, texture, depression, and convex processing, etc., and finally a physical three-dimensional model of the vase is obtained.
在本实施例中,通过采用包括上述步骤S410至步骤S440的三维模型传输方法,首先服务器端对实物图像集合进行解析处理,得到图形信息,根据图形信息从实物模型库中获取实物图像元素信息,接着利用实物图像元素信息构建实物图像底图,最后对实物图像底图进行建模处理,得到实物三维模型,因此,本申请实施例能够通过图形信息从实物模型库中获取实物图像元素信息,利用实物图像元素信息构建实物图像底图,以达到缩短建模时间和提高建模效率的目的。In this embodiment, by adopting the three-dimensional model transmission method including the above-mentioned steps S410 to S440, first, the server side parses the physical image collection to obtain graphic information, and obtains the physical image element information from the physical model library according to the graphic information. Then, the physical image element information is used to construct the physical image base map, and finally the physical image base map is modeled to obtain the physical three-dimensional model. Therefore, the embodiment of the present application can obtain the physical image element information from the physical model library through the graphic information, using The physical image element information is used to construct the physical image base map to shorten the modeling time and improve the modeling efficiency.
在一实施例中,如图5所示,对步骤S140进行进一步的说明,该步骤S140可以包括但不限于有步骤S510、步骤S520和步骤S530。In one embodiment, as shown in FIG. 5 , step S140 is further described. This step S140 may include but is not limited to step S510, step S520 and step S530.
步骤S510:对环境图像集合进行解析处理,得到标签信息。Step S510: Analyze the environment image set to obtain label information.
在一实施方式中,标签信息可以包括环境外观的材料属性信息、颜色属性信息、光源属性信息等,在此不做具体限制。In one embodiment, the tag information may include material attribute information, color attribute information, light source attribute information, etc. of environmental appearance, which are not specifically limited here.
步骤S520:根据标签信息从环境模型库中获取环境图像元素信息。Step S520: Obtain environmental image element information from the environment model library according to the label information.
在一实施方式中,环境图像元素信息可以包括环境外观的材料属性信息、颜色属性信息、光源属性信息等的索引信息,在此不做具体限制。In one embodiment, the environment image element information may include index information such as material attribute information, color attribute information, light source attribute information, etc. of the environment appearance, which is not specifically limited here.
步骤S530:利用环境图像元素信息对环境图像集合进行建模处理,得到环境三维模型。Step S530: Use the environment image element information to perform modeling processing on the environment image set to obtain a three-dimensional environment model.
在一实施方式中,利用环境图像元素信息对环境图像集合进行建模处理,可以包括材质贴面、调光、纹理、凹陷、凸起处理、塌陷处理等,使得模型具有立体感,在此不做具体限制。In one embodiment, the environment image element information is used to model the environment image collection, which may include material veneer, light adjustment, texture, depression, bulge processing, collapse processing, etc., so that the model has a three-dimensional sense. This is not the case here. Make specific restrictions.
在本实施例中,通过采用包括上述步骤S510至步骤S530的三维模型传输方法,首先服务器端对环境图像集合进行解析处理,得到标签信息,然后根据标签信息从环境模型库中获取环境图像元素信息,最后利用环境图像元素信息对环境图像集合进行建模处理,得到环境三维模型,因此,本申请实施例通过标签信息从环境模型库中获取环境图像元素信息,利用环境图像元素信息对环境图像集合进行建模处理,以达到缩短建模时间和提高建模效率的目的。In this embodiment, by adopting the three-dimensional model transmission method including the above steps S510 to S530, first the server parses the environment image collection to obtain label information, and then obtains environmental image element information from the environment model library according to the label information. , and finally the environment image element information is used to model the environment image collection to obtain the environment three-dimensional model. Therefore, the embodiment of the present application obtains the environment image element information from the environment model library through the tag information, and uses the environment image element information to model the environment image collection. Carry out modeling processing to achieve the purpose of shortening modeling time and improving modeling efficiency.
在一实施例中,如图6所示,该三维模型传输方法还可以包括但不限于有步骤S610和步骤S620。In an embodiment, as shown in Figure 6, the three-dimensional model transmission method may also include but is not limited to step S610 and step S620.
步骤S610:根据实物三维模型从环境模型库中确定目标环境三维模型。Step S610: Determine the target environment three-dimensional model from the environment model library according to the physical three-dimensional model.
在一实施方式中,环境模型库中可以包括多个预设环境模型,该多个预设环境模型可以是由模型制作人员使用三维建模工具(比如3D Studio MAX等建模工具,其中,3D Studio MAX是基于电脑操作系统的三维动画渲染和制作软件)对已有模型数据进行修改后,利用修改后的模型数据进行建模处理后得到的,也可以是已有的模型,在此不做具体限制。In one embodiment, the environment model library may include multiple preset environment models. The multiple preset environment models may be created by model makers using three-dimensional modeling tools (such as 3D Studio MAX and other modeling tools, where 3D Studio MAX is a three-dimensional animation rendering and production software based on a computer operating system). It is obtained by modifying the existing model data and using the modified model data for modeling processing. It can also be an existing model, which will not be done here. Specific restrictions.
在一实施例中,可以将实物三维模型与环境模型库中预设环境模型进行适配处理,确定目标环境三维模型,而不需要对环境图像集合进行建模处理,缩短了建模时间,本申请实施例对此不做具体限制。In one embodiment, the physical three-dimensional model can be adapted to the preset environment model in the environment model library to determine the target environment three-dimensional model without modeling the environment image collection, which shortens the modeling time. The application examples do not specifically limit this.
步骤S620:将实物三维模型与目标环境三维模型进行嵌套处理,生成第一三维场景模型。Step S620: Nest the physical three-dimensional model and the target environment three-dimensional model to generate a first three-dimensional scene model.
本实施例中,通过采用包括上述步骤S610至步骤S620的三维模型传输方法,首先服务器端根据实物三维模型从环境模型库中确定目标环境三维模型,最后将实物三维模型与目标环境三维模型进行嵌套处理,生成第一三维场景模型,即是说,可以只对实物图像集合进行建模处理,得到实物三维模型,只根据实物三维模型从环境模型库中确定目标环境三维模型,将实物三维模型与目标环境三维模型进行嵌套处理,生成第一三维场景模型,而不需要对环境图像集合进行建模处理,缩短了建模时间,同时可以将客户端或者设计端所处场景中的实物通过三维模型的形式与环境模型库中的预设环境三维模型进行适配,达到远程设计的效果,满足了用户的多样化需求,提高用户体验。In this embodiment, by adopting the three-dimensional model transmission method including the above steps S610 to step S620, first the server determines the target environment three-dimensional model from the environment model library based on the physical three-dimensional model, and finally embeds the physical three-dimensional model and the target environment three-dimensional model. A set of processes is performed to generate the first three-dimensional scene model. That is to say, you can only perform modeling processing on the physical image collection to obtain the physical three-dimensional model. You can only determine the target environment three-dimensional model from the environment model library based on the physical three-dimensional model, and convert the physical three-dimensional model to the first three-dimensional scene model. It is nested with the target environment 3D model to generate the first 3D scene model without the need to model the environment image collection, which shortens the modeling time. At the same time, the actual objects in the scene where the client or the design end is located can be passed through The form of the three-dimensional model is adapted to the preset environmental three-dimensional model in the environment model library to achieve the effect of remote design, meet the diverse needs of users, and improve user experience.
在一实施例中,当对客厅进行家居布局设计时,客户端可以将包含沙发、冰箱、座椅等实物图像集合发送给服务器端,服务器端对实物图像集合进行建模处理,得到多个实物三维 模型,接着服务器端根据该多个实物三维模型从环境模型库中确定目标环境三维模型,将实物三维模型与目标环境三维模型进行嵌套处理,生成第一三维场景模型,因此,本申请实施例不需要对环境图像集合进行建模处理,缩短了建模时间。In one embodiment, when designing the home layout of the living room, the client can send a collection of physical objects including sofas, refrigerators, seats, etc. to the server, and the server performs modeling processing on the physical image collection to obtain multiple physical objects. three dimensional model, and then the server determines the target environment three-dimensional model from the environment model library based on the multiple physical three-dimensional models, performs nesting processing on the physical three-dimensional model and the target environment three-dimensional model, and generates the first three-dimensional scene model. Therefore, the embodiment of the present application There is no need to perform modeling processing on the environment image collection, which shortens the modeling time.
在一实施例中,对步骤S150进行进一步的说明,该步骤S150可以包括但不限于有以下步骤:In an embodiment, step S150 is further described. Step S150 may include but is not limited to the following steps:
将实物三维模型与环境三维模型进行嵌套处理,生成第一三维场景模型。The physical three-dimensional model and the environmental three-dimensional model are nested to generate a first three-dimensional scene model.
本实施例中,服务器端可以获取来自设计端的多个视频图像,视频图像由设计端通过对增强现实视频通话的视频流进行抽帧处理而获得,增强现实视频通话由设计端与客户端建立,接着对多个视频图像进行分割处理,得到实物图像集合和环境图像集合,然后对实物图像集合进行建模处理,得到实物三维模型,并且对环境图像集合进行解析处理,得到图形信息,然后将实物三维模型与环境三维模型进行嵌套处理,生成第一三维场景模型,因此,本申请实施例能够通过对用户所处场景中的环境和实物进行建模,并将实物三维模型与环境三维模型进行嵌套处理得到第一三维场景模型,还原真实场景,达到远程定制的效果。In this embodiment, the server side can obtain multiple video images from the design side. The video images are obtained by the design side by performing frame extraction processing on the video stream of the augmented reality video call. The augmented reality video call is established by the design side and the client. Then multiple video images are segmented to obtain a collection of physical images and a collection of environmental images. Then the physical image collection is modeled to obtain a three-dimensional model of the physical object. The environmental image collection is analyzed and processed to obtain graphic information, and then the physical object is The three-dimensional model and the environment three-dimensional model are nested to generate the first three-dimensional scene model. Therefore, the embodiment of the present application can model the environment and physical objects in the scene where the user is located, and combine the physical three-dimensional model with the environmental three-dimensional model. The first three-dimensional scene model is obtained through nesting processing, restoring the real scene and achieving the effect of remote customization.
在一实施例中,当服务器端可以获取来自设计端的多个视频图像,对多个视频图像进行分割处理,得到实物图像集合和环境图像集合,然后对实物图像集合进行建模处理,得到实物三维模型,并且对环境图像集合进行解析处理,得到标签信息,根据标签信息与环境模型库中的预设环境模型进行适配,当标签信息与预设环境模型不适配,对环境图像集合进行建模处理,得到环境三维模型,将实物三维模型与环境三维模型进行嵌套处理得到第一三维场景模型。另外,可以将该环境三维模型存储到环境模型库中,本申请实施对此不做具体限制。In one embodiment, when the server side can obtain multiple video images from the design side, it performs segmentation processing on the multiple video images to obtain a set of physical images and a set of environmental images, and then performs modeling processing on the set of physical images to obtain a three-dimensional physical object. model, and parses the environment image collection to obtain label information, and adapts the label information to the preset environment model in the environment model library. When the label information does not match the preset environment model, the environment image collection is constructed. Model processing is performed to obtain an environment three-dimensional model, and the physical three-dimensional model and the environment three-dimensional model are nested to obtain the first three-dimensional scene model. In addition, the environment three-dimensional model can be stored in the environment model library, and this application does not impose specific restrictions on this.
在一实施例中,首先服务器端获取客户端发送的环境图像集合,并且对环境图像集合进行建模处理,得到环境三维模型,然后服务器端获取来自设计端的多个视频图像,视频图像由设计端通过对增强现实视频通话的视频流进行抽帧处理而获得,增强现实视频通话由设计端与客户端建立,接着对多个视频图像进行分割处理,得到实物图像集合,然后对实物图像集合进行建模处理,得到实物三维模型,根据实物三维模型和环境三维模型生成第一三维场景模型,最后将第一三维场景模型发送至客户端,因此,本申请实施例通过对客户端发送的环境图像集合进行建模处理,以满足客户需求,达到远程定制的效果。In one embodiment, first, the server acquires the set of environment images sent by the client, and performs modeling processing on the set of environment images to obtain a three-dimensional model of the environment. Then, the server acquires multiple video images from the design end, and the video images are generated by the design end. It is obtained by extracting frames from the video stream of the augmented reality video call. The augmented reality video call is established by the design end and the client. Then multiple video images are segmented to obtain a set of physical images, and then the set of physical images is constructed. Model processing is performed to obtain the physical three-dimensional model, the first three-dimensional scene model is generated according to the physical three-dimensional model and the environmental three-dimensional model, and finally the first three-dimensional scene model is sent to the client. Therefore, the embodiment of the present application processes the environment image collection sent by the client. Carry out modeling processing to meet customer needs and achieve remote customization effects.
另外,图7是本申请另一个实施例提供的一种三维模型传输方法,该三维模型传输方法可以包括但不限于步骤S710和步骤S720。In addition, FIG. 7 is a three-dimensional model transmission method provided by another embodiment of the present application. The three-dimensional model transmission method may include but is not limited to step S710 and step S720.
步骤S710:与设计端建立增强现实视频通话。Step S710: Establish an augmented reality video call with the design end.
在一实施方式中,设计端向客户端发送视频入会邀请,客户端确认接受后,经过服务器端的排队系统和服务器端的鉴权系统的确认,设计端才可以与客户端建立增强现实视频通话;或者,客户端向设计端发送视频入会邀请,设计端确认接受后,经过服务器端的排队系统和服务器端的鉴权系统的确认,客户端才可以与设计端建立增强现实视频通话,在此不做具体限制。In one implementation, the design end sends a video conference invitation to the client. After the client confirms acceptance, the design end can establish an augmented reality video call with the client only after confirmation by the server-side queuing system and the server-side authentication system; or , the client sends a video conference invitation to the design end. After the design end confirms acceptance, the client can establish an augmented reality video call with the design end after being confirmed by the server-side queuing system and the server-side authentication system. There are no specific restrictions here. .
步骤S720:接收服务器端发送的第一三维场景模型。Step S720: Receive the first three-dimensional scene model sent by the server.
在一实施方式中,第一三维场景模型由服务器端根据实物三维模型和环境三维模型而获得,实物三维模型由服务器端通过对实物图像集合进行建模处理而获得,环境三维模型由服务器端通过对环境图像集合进行建模处理而获得,实物图像集合和环境图像集合均由服务器端通过对多个视频图像进行分割处理而获得,视频图像由设计端通过对增强现实视频通话的视频流进行抽帧处理而获得。In one embodiment, the first three-dimensional scene model is obtained by the server based on the physical three-dimensional model and the environmental three-dimensional model. The physical three-dimensional model is obtained by the server through modeling processing of the physical image collection, and the environmental three-dimensional model is obtained by the server. The environmental image collection is obtained through modeling processing. Both the physical image collection and the environmental image collection are obtained by segmenting multiple video images on the server side. The video images are extracted by the design end by extracting the video stream of the augmented reality video call. obtained through frame processing.
在一实施方式中,对多个视频图像进行分割处理,可以有很多实施方式,比如,可以是采用边缘轮廓扫描的方式对多个视频图像进行分割处理,得到实物图像集合和环境图像集合;或者,可以利用深度学习算法对多个视频图像中的实物图像和环境图像进行识别处理,并根据构图规则在该多个视频图像中分别确定与识别出的实物图像对应的分割区域和环境图像对应的分割区域,再根据该实物图像对应的分割区域和环境图像对应的分割区域,得到的实物图像集合和环境图像集合,其中,该构图规则可以包括将实物图像或者环境图像在分割区域 中所处的位置、所占的面积进行设定的规则,在此不做具体限制。In one embodiment, there are many implementation methods for segmenting multiple video images. For example, edge contour scanning may be used to segment multiple video images to obtain a set of physical images and a set of environmental images; or , the deep learning algorithm can be used to identify and process the physical object images and environmental images in multiple video images, and the segmented areas corresponding to the recognized physical images and the environmental images corresponding to the multiple video images are determined respectively according to the composition rules. segmentation area, and then obtain the physical image set and the environmental image set based on the segmentation area corresponding to the physical image and the segmentation area corresponding to the environmental image, wherein the composition rule may include placing the physical image or the environmental image in the segmentation area The rules for setting the location and area occupied are not specifically limited here.
可以理解的是,由于视频的本质是一帧帧连续的图片,所以设计端可以从视频中截取一帧清晰图像,也可以截取多张图像集,即设计端通过对增强现实视频通话的视频流进行抽帧处理,得到多个视频图像。It is understandable that since the essence of video is continuous pictures frame by frame, the design end can intercept a clear frame of image from the video, or it can intercept multiple image sets, that is, the design end can intercept the video stream of the augmented reality video call. Perform frame extraction processing to obtain multiple video images.
在一实施方式中,在增强现实视频通话的过程中,可以支持音视频切换,比如,客户端在增强现实视频通话的过程中,客户端可以切换到应用A的播放界面,并将增强现实视频通话置为后台程序,使增强现实视频通话在后台运行,其中,应用A可以是音频,也可以是视频;又如,客户端在与设计端进行增强现实视频通话的过程中,客户端可以将视频通话切换成语音通话,在此不做具体限制。In one embodiment, during the augmented reality video call, audio and video switching can be supported. For example, during the augmented reality video call, the client can switch to the playback interface of application A and display the augmented reality video. The call is set as a background program so that the augmented reality video call runs in the background. Application A can be audio or video. For another example, when the client is having an augmented reality video call with the designer, the client can There are no specific restrictions on switching a video call to a voice call.
在一实施方式中,增强现实视频通话的视频流采用YUV编码和H.264视频协议,其中,YUV为一种颜色编码格式,“Y”表示明亮度,即灰阶值;“U”和“V”表示色度,“U”和“V”均可以描述影像色彩及饱和度,可以用于指定像素的颜色,其中,H.264视频协议为一种数字视频压缩编码标准。In one implementation, the video stream of the augmented reality video call adopts YUV encoding and H.264 video protocol, where YUV is a color encoding format, "Y" represents brightness, that is, grayscale value; "U" and " V" represents chroma. Both "U" and "V" can describe image color and saturation, and can be used to specify the color of pixels. Among them, the H.264 video protocol is a digital video compression encoding standard.
在一实施方式中,增强现实视频通话的视频流可以是客户端侧的视频流,也可以是设计端侧的视频流,也可以是客户端侧的视频流和用户端侧的视频流,可以根据实际情况获取,在此不做具体限制。比如,在设计端与客户端建立增强现实视频通话后,设计端可以对设计端侧的视频流进行抽帧处理,得到多个视频图像;或者,在设计端与客户端建立增强现实视频通话后,设计端可以对客户端侧的视频流进行抽帧处理,得到多个视频图像;又或者,在设计端与客户端建立增强现实视频通话后,设计端可以分别对设计端侧的视频流和客户端侧的视频流进行抽帧处理,得到多个视频图像。In one embodiment, the video stream of the augmented reality video call can be a video stream on the client side, a video stream on the design side, or a video stream on the client side and a video stream on the user side. It is obtained based on the actual situation and there are no specific restrictions here. For example, after the design end and the client establish an augmented reality video call, the design end can extract frames from the video stream on the design end to obtain multiple video images; or, after the design end and the client establish an augmented reality video call, , the design end can perform frame extraction processing on the video stream on the client side to obtain multiple video images; or, after the design end and the client establish an augmented reality video call, the design end can process the video stream on the design end side and the client side respectively. The video stream on the client side undergoes frame extraction processing to obtain multiple video images.
在本实施例中,通过采用包括上述步骤S710至S720的三维模型传输方法,首先客户端与设计端建立增强现实视频通话,然后客户端接收服务器端发送的第一三维场景模型,其中,该第一三维场景模型由服务器端根据实物三维模型和环境三维模型而获得,实物三维模型由服务器端通过对实物图像集合进行建模处理而获得,环境三维模型由服务器端通过对环境图像集合进行建模处理而获得,实物图像集合和环境图像集合均由服务器端通过对多个视频图像进行分割处理而获得,视频图像由设计端通过对增强现实视频通话的视频流进行抽帧处理而获得,即是说,在设计端与客户端建立的增强现实视频通话的场景下,设计端对该增强现实视频通话的视频流进行抽帧处理得到多个视频图像,服务器端通过对该多个视频图像进行处理,最终得到第一三维场景模型,最后客户端接收服务器端发送的第一三维场景模型,因此,本申请实施例能够在增强现实场景下实现三维模型的传输。In this embodiment, by adopting the three-dimensional model transmission method including the above steps S710 to S720, first the client establishes an augmented reality video call with the design terminal, and then the client receives the first three-dimensional scene model sent by the server, wherein the third A three-dimensional scene model is obtained by the server side based on the physical three-dimensional model and the environmental three-dimensional model. The physical three-dimensional model is obtained by the server side by modeling a collection of physical objects. The environmental three-dimensional model is obtained by the server side by modeling a collection of environmental images. The physical image collection and the environment image collection are obtained by segmenting multiple video images on the server side, and the video images are obtained by the design end by extracting frames from the video stream of the augmented reality video call, that is, Say, in the scenario of an augmented reality video call established between the design end and the client, the design end performs frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images, and the server end processes the multiple video images. , finally obtain the first three-dimensional scene model, and finally the client receives the first three-dimensional scene model sent by the server. Therefore, the embodiment of the present application can realize the transmission of the three-dimensional model in the augmented reality scene.
在一实施例中,如图8所示,该三维模型传输方法可以包括但不限于步骤S810和步骤S820。In an embodiment, as shown in FIG. 8 , the three-dimensional model transmission method may include but is not limited to step S810 and step S820.
步骤S810:对第一三维场景模型进行虚拟标注处理,得到标注信息。Step S810: Perform virtual annotation processing on the first three-dimensional scene model to obtain annotation information.
在一实施方式中,标注信息可以包括坐标信息、文字信息、标注过的模型等,在此不做具体限制。In one implementation, the annotation information may include coordinate information, text information, annotated models, etc., and is not specifically limited here.
在一实施方式中,服务器端可以将第一三维场景模型发送给设计端,设计端对该第一三维场景模型进行存储处理。当客户端对第一三维场景模型进行虚拟标注处理时,客户端可以实时将该标注信息发送给设计端,比如,客户端可以对第一三维场景模型进行视角切换,当第一三维场景模型为一个包括卧室、客厅和卫生间的房子时,客户端可以将第一三维场景模型切换到卧室区域,并对该卧室区域进行虚拟标注处理,也可以将该第一三维场景模型切换到客厅区域,并对该客厅区域进行虚拟标注处理,还可以将该第一三维场景模型切换到卫生间区域,并对该卫生间区域进行虚拟标注处理,最后将所有区域的标注信息发送给设计端(或者服务器端),设计端可以实时接收该所有区域的标注信息,并根据该标注信息对存储在设计端中的第一三维场景模型进行修改设计,实现对三维模型的远程修改,提高用户体验,在此不做具体限制。In one implementation, the server may send the first three-dimensional scene model to the design end, and the design end may store and process the first three-dimensional scene model. When the client performs virtual annotation processing on the first three-dimensional scene model, the client can send the annotation information to the design end in real time. For example, the client can switch the perspective of the first three-dimensional scene model. When the first three-dimensional scene model is When a house includes a bedroom, a living room and a bathroom, the client can switch the first three-dimensional scene model to the bedroom area and perform virtual annotation processing on the bedroom area, or it can switch the first three-dimensional scene model to the living room area and Perform virtual annotation processing on the living room area, switch the first three-dimensional scene model to the bathroom area, perform virtual annotation processing on the bathroom area, and finally send the annotation information of all areas to the design end (or server end). The design end can receive the annotation information of all areas in real time, and modify the design of the first three-dimensional scene model stored in the design end based on the annotation information to realize remote modification of the three-dimensional model and improve the user experience. This will not be detailed here. limit.
步骤S820:将标注信息发送至服务器端。 Step S820: Send the annotation information to the server.
在一实施方式中,客户端也可以将该标准信息发送给设计端,设计端根据标准信息对存储在设计端中的第一三维场景模型进行修改设计,由于第一三维场景模型中的各个模块可能由不同的设计师负责,即多方设计师对第一三维场景模型进行并发修改,因此,可能会出现并发冲突。为规避并发冲突的问题,设计端中的多名设计师可以分别根据自身的设计副本,对与模块对应的源数据文件进行修改,最后设计端将所有修改后的源数据文件提交给服务器端,服务器端对该源数据文件进行收集后再统一交付给用户,因此,本申请实施例这样直接根据标注信息对第一三维场景模型进行修改,减少沟通成本,实现快速交付。In one implementation, the client can also send the standard information to the design end, and the design end modifies the design of the first three-dimensional scene model stored in the design end based on the standard information. Since each module in the first three-dimensional scene model Different designers may be responsible for it, that is, multiple designers may make concurrent modifications to the first three-dimensional scene model. Therefore, concurrency conflicts may occur. In order to avoid the problem of concurrency conflicts, multiple designers in the design end can modify the source data files corresponding to the modules based on their own design copies. Finally, the design end submits all modified source data files to the server end. The server collects the source data files and then delivers them to the user in a unified manner. Therefore, in this embodiment of the present application, the first three-dimensional scene model is directly modified based on the annotation information, thereby reducing communication costs and achieving rapid delivery.
在一实施方式中,将标注信息保存在相关文件中,打包成数据包发送至服务器端,服务器端接收到该数据包后,对该数据包进行解析处理,得到标注信息,在此不做具体限制。In one implementation, the annotation information is saved in a relevant file, packaged into a data packet and sent to the server. After receiving the data packet, the server parses the data packet to obtain the annotation information, which will not be detailed here. limit.
在本实施例中,通过采用包括上述步骤S810至S820的三维模型传输方法,首先客户端可以对第一三维场景模型进行虚拟标注处理,得到标注信息,并将标注信息发送至服务器端,以使服务器端根据标注信息在第一三维场景模型中确定标注区域,在标注区域上嵌套标注信息,得到第二三维场景模型,最后将第一三维场景模型和第二三维场景模型进行叠加处理,得到第三三维场景模型,即是说,客户端可以将标注信息反馈给服务器,通过服务器根据标注信息对第一三维场景模型进行修改,减少沟通成本,同时有利于满足用户的需求,提高用户体验。In this embodiment, by adopting the three-dimensional model transmission method including the above steps S810 to S820, first, the client can perform virtual annotation processing on the first three-dimensional scene model, obtain annotation information, and send the annotation information to the server, so that the The server determines the annotation area in the first three-dimensional scene model based on the annotation information, nests the annotation information on the annotation area, and obtains the second three-dimensional scene model. Finally, the first three-dimensional scene model and the second three-dimensional scene model are superimposed to obtain The third three-dimensional scene model means that the client can feed back the annotation information to the server, and the server can modify the first three-dimensional scene model based on the annotation information, thereby reducing communication costs, while also helping to meet user needs and improve user experience.
在一实施例中,在旅游实景讲解的场景下,用户通过在AR设备中对建筑物B进行标记,并将该建筑物B通过AR设备发送给服务器端,服务器端对实物图像集合进行解析处理,得到图形信息,根据图形信息从实物模型库中获取实物图像元素信息,接着利用实物图像元素信息构建实物图像底图,对实物图像底图进行建模处理,得到实物三维模型,最后将该实物三维模型发送至客户端,并且用户还通过AR设备将该建筑物B发送给设计端,设计端根据该建筑物B从数据存储模块中获取与建筑物B相关的历史资料,并将该历史资料发送给AR设备,其中,设计端可以通过语音、视频或者文字等形式将该历史资料发送给AR设备,除此之外,用户还可以通过语音控制或外设输入的方式对AR设备中的实物三维模型进行放大或者缩小,以及切换对AR设备中的实物三维模型的观察视角,以便查看模型细节,其中,AR设备为客户端,在此不做具体限制。In one embodiment, in the scenario of real-life tourism explanation, the user marks building B in the AR device, and sends the building B to the server through the AR device, and the server parses the collection of physical images. , obtain the graphic information, obtain the physical image element information from the physical model library according to the graphic information, then use the physical image element information to construct the physical image base map, perform modeling processing on the physical image base map, obtain the physical three-dimensional model, and finally convert the physical object The three-dimensional model is sent to the client, and the user also sends the building B to the design end through the AR device. The design end obtains the historical data related to building B from the data storage module based on the building B and saves the historical data. Send it to the AR device. The design end can send the historical data to the AR device through voice, video or text. In addition, the user can also control the physical objects in the AR device through voice control or peripheral input. The three-dimensional model is enlarged or reduced, and the viewing angle of the physical three-dimensional model in the AR device is switched to view the model details. The AR device is the client and is not specifically limited here.
针对上述实施例所提供的三维模型传输方法,下面以一实施例进行详细的描述:Regarding the three-dimensional model transmission method provided by the above embodiments, a detailed description is given below using an embodiment:
在一实施例中,参考图9,首先,设计端和客户端建立增强现实视频通话,接着设计端对增强现实视频通话的视频流进行抽帧处理,得到多个视频图像,接着对多个视频图像进行分割处理得到实物图像集合和环境图像集合,紧接着分别对实物图像集合和环境图像集合进行解析处理,其中,针对实物图像集合的解析处理,可以对实物图像集合进行解析处理得到图形信息,根据图形信息从实物模型库中获取实物图像元素信息,利用实物图像元素信息构建实物图像底图,对实物图像底图进行建模处理,得到实物三维模型;针对环境图像集合的解析处理,可以对环境图像集合进行解析处理得到标签信息,根据标签信息从环境模型库中获取环境图像元素信息,利用环境图像元素信息对环境图像集合进行建模处理得到环境三维模型,然后根据实物三维模型和环境三维模型生成第一三维场景模型,将第一三维场景模型发送至客户端,客户端对第一三维场景模型进行三维渲染处理;或者,服务器对第一三维场景模型进行三维渲染处理得到渲染后的第一三维场景模型,之后,将渲染后的第一三维场景模型发送至客户端,客户端再次对渲染后的第一三维场景模型进行三维渲染处理得到再次渲染后的第一三维场景模型,之后,客户端对渲染后的第一三维场景模型或者再次渲染后的第一三维场景模型进行虚拟标注处理,得到标注信息,客户端将该标注信息对该渲染后的第一三维场景模型或者再次渲染后的第一三维场景模型进行并发修改,即设计端中的多名设计师根据自身的设计副本,对与模块对应的源数据文件进行修改,最后设计端将所有修改后的源数据文件提交给服务器端,服务器端将该源数据文件统一发送至客户端,实现远程交付。In one embodiment, referring to Figure 9, first, the design end and the client establish an augmented reality video call, and then the design end performs frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images, and then processes the multiple videos. The image is segmented to obtain a set of physical images and a set of environmental images, and then the set of physical images and the set of environmental images are analyzed and processed respectively. Among them, for the analysis and processing of the set of physical images, the set of physical images can be analyzed and processed to obtain graphic information. Obtain the physical image element information from the physical model library according to the graphic information, use the physical image element information to construct the physical image base map, perform modeling processing on the physical image base map, and obtain the physical three-dimensional model; for the analysis and processing of the environmental image collection, you can The environmental image collection is analyzed and processed to obtain label information. The environmental image element information is obtained from the environmental model library according to the label information. The environmental image element information is used to model the environmental image collection to obtain an environmental three-dimensional model. Then, the environmental three-dimensional model is obtained based on the physical three-dimensional model and the environmental three-dimensional model. The model generates a first three-dimensional scene model, and sends the first three-dimensional scene model to the client, and the client performs three-dimensional rendering processing on the first three-dimensional scene model; or, the server performs three-dimensional rendering processing on the first three-dimensional scene model to obtain the rendered third A three-dimensional scene model, and then the rendered first three-dimensional scene model is sent to the client. The client performs three-dimensional rendering processing on the rendered first three-dimensional scene model again to obtain the re-rendered first three-dimensional scene model. After that, The client performs virtual annotation processing on the rendered first three-dimensional scene model or the re-rendered first three-dimensional scene model to obtain annotation information, and the client applies the annotation information to the rendered first three-dimensional scene model or the re-rendered first three-dimensional scene model. The first three-dimensional scene model is modified concurrently, that is, multiple designers in the design end modify the source data files corresponding to the module based on their own design copies, and finally the design end submits all modified source data files to the server The server side sends the source data files to the client in a unified manner to achieve remote delivery.
另外,参照图10,本申请的一个实施例还提供了一种三维模型传输装置,该三维模型传输装置包括设计端100、客户端300和服务器端200。 In addition, referring to FIG. 10 , an embodiment of the present application also provides a three-dimensional model transmission device. The three-dimensional model transmission device includes a design terminal 100, a client 300, and a server terminal 200.
其中,设计端100包括第一显示模块101、第一三维渲染模块102、第一音视频处理模块103、数据存储模块104和坐席集成接口105。其中,第一显示模块101可以用于显示渲染后的第一三维场景模型或者第三三维场景模型,或者,用于显示再次渲染后的第一三维场景模型或者第三三维场景模型,其中,显示模块包括电脑显示屏或AR屏幕等;第一三维渲染模块102可以用于对第一三维场景模型或者第三三维场景模型进行三维渲染处理,得到渲染后的第一三维场景模型或者渲染后的第三三维场景模型,也可以用于对渲染后的第一三维场景模型或者渲染后的第三三维场景模型进行再次三维渲染处理,得到再次渲染后的第一三维场景模型或者再次渲染后的第三三维场景模型;第一音视频处理模块103可以用于接收对坐席集成接口105发送的增强现实视频通话的视频流,并对增强现实视频通话的视频流进行抽帧处理,得到多个视频图像,还可以用于接收来自坐席集成接口105的第一三维场景模型或者第三三维场景模型,或者接收来自坐席集成接口105的渲染后的第一三维场景模型或者渲染后的第三三维场景模型;数据存储模块104可以用于存储来自坐席集成接口105的增强现实视频通话的视频流、标注信息、模型相关数据和第一三维场景模型、第三三维场景模型、渲染后的第一三维场景模型和渲染后的第三三维场景模型等;坐席集成接口105可以用于获取第一音视频处理模块103中的多个视频图像,并对该多个视频图像发送至服务器端200,还可以获取服务器端200的第一三维场景模型、第三三维场景模型、渲染后的第一三维场景模型和渲染后的第三三维场景模型等。Among them, the design end 100 includes a first display module 101, a first three-dimensional rendering module 102, a first audio and video processing module 103, a data storage module 104 and an agent integration interface 105. The first display module 101 may be used to display the rendered first three-dimensional scene model or the third three-dimensional scene model, or to display the re-rendered first three-dimensional scene model or the third three-dimensional scene model, wherein the display The module includes a computer display screen or an AR screen, etc.; the first three-dimensional rendering module 102 can be used to perform three-dimensional rendering processing on the first three-dimensional scene model or the third three-dimensional scene model to obtain the rendered first three-dimensional scene model or the rendered third three-dimensional scene model. The three-dimensional scene model can also be used to perform three-dimensional rendering processing on the rendered first three-dimensional scene model or the rendered third three-dimensional scene model to obtain the first three-dimensional scene model after re-rendering or the third three-dimensional scene model after re-rendering. Three-dimensional scene model; the first audio and video processing module 103 can be used to receive the video stream of the augmented reality video call sent to the agent integration interface 105, and perform frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images, It can also be used to receive the first three-dimensional scene model or the third three-dimensional scene model from the agent integration interface 105, or to receive the rendered first three-dimensional scene model or the rendered third three-dimensional scene model from the agent integration interface 105; data The storage module 104 may be used to store the video stream of the augmented reality video call from the agent integration interface 105, annotation information, model-related data, the first three-dimensional scene model, the third three-dimensional scene model, the rendered first three-dimensional scene model and the rendering. The third three-dimensional scene model, etc.; the agent integration interface 105 can be used to obtain multiple video images in the first audio and video processing module 103, and send the multiple video images to the server 200, and can also obtain the server 200 The first three-dimensional scene model, the third three-dimensional scene model, the rendered first three-dimensional scene model, the rendered third three-dimensional scene model, etc.
在一实施方式中,第一音视频处理模块103可以支持音视频编码格式;可以调用数据存储模块104;可以根据优先级对视频入会邀请进行排序,并且可以对视频入会邀请选择视频应答或者音频应答;同时支持多流的媒体协商模式;可以将第一三维场景模型或者第三三维场景模型传输给第一三维渲染模块102。In one embodiment, the first audio and video processing module 103 can support audio and video encoding formats; can call the data storage module 104; can sort video conference invitations according to priority, and can select video responses or audio responses for video conference invitations. ;Support multi-stream media negotiation mode simultaneously; The first three-dimensional scene model or the third three-dimensional scene model can be transmitted to the first three-dimensional rendering module 102.
在一实施方式中,坐席集成接口105可用于接收服务器端200的接收消息处理模块203的访问请求,并对该访问请求进行解析,还可以获取第一音视频处理模块103中的相关信息进行打包得到数据包,将该数据包发送至服务器端200,在此不做具体限制。In one embodiment, the agent integration interface 105 can be used to receive an access request from the message processing module 203 of the server 200, parse the access request, and also obtain relevant information from the first audio and video processing module 103 for packaging. Obtain the data packet and send the data packet to the server 200, which is not specifically limited here.
服务器端200包括文件存储器201、消息处理模块203、消息缓存模块204、数据流模块205、图像解析模块209、模型适配模块210、三维模型构造器211、模型存储模块208和第二三维渲染模块206。其中,文件存储器201可以用于存储来自坐席集成接口105的多个视频图像;消息处理模块203可以用于对视频入会邀请进行鉴权确认;消息缓存模块204可以用于接收消息处理模块203的用户信息以及坐席地址等数据;数据流模块205可以用于对来自设计端100的坐席集成接口105的增强现实视频通话的视频流进行传输,并可用于发送视频入会邀请以及退出增强现实视频通话;图像解析模块209可以用于对实物图像集合进行解析处理得到图形信息,并且对环境图像集合进行解析处理得到标签信息;模型适配模块210可以用于利用实物图像元素信息构建实物图像底图,并对实物图像底图进行建模处理,得到实物三维模型;模型存储模块208可以用于根据标签信息从环境模型库中获取环境图像元素信息,并且还可以用于根据图形信息从实物模型库中获取实物图像元素信息;三维模型构造器211可以用于利用环境图像元素信息对环境图像集合进行建模处理,得到环境三维模型;第二三维渲染模块206可以用于将来自三维模型构造器211的第一三维场景模型或者第三三维场景模型进行三维渲染处理,得到渲染后的第一三维场景模型或者第三三维场景模型。The server 200 includes a file storage 201, a message processing module 203, a message cache module 204, a data flow module 205, an image analysis module 209, a model adaptation module 210, a three-dimensional model constructor 211, a model storage module 208 and a second three-dimensional rendering module. 206. Among them, the file storage 201 can be used to store multiple video images from the agent integration interface 105; the message processing module 203 can be used to authenticate and confirm the video conference invitation; the message caching module 204 can be used to receive users of the message processing module 203 Information and data such as agent addresses; the data flow module 205 can be used to transmit the video stream of the augmented reality video call from the agent integration interface 105 of the design end 100, and can be used to send video invitations to join the meeting and exit the augmented reality video call; image The parsing module 209 can be used to parse the physical image set to obtain graphic information, and to parse the environmental image set to obtain label information; the model adaptation module 210 can be used to construct the physical image base map using the physical image element information, and The physical image base map is subjected to modeling processing to obtain a three-dimensional model of the physical object; the model storage module 208 can be used to obtain environmental image element information from the environmental model library according to the label information, and can also be used to obtain the physical object from the physical model library according to the graphic information. Image element information; the three-dimensional model constructor 211 can be used to perform modeling processing on the environment image set using the environment image element information to obtain an environment three-dimensional model; the second three-dimensional rendering module 206 can be used to convert the first three-dimensional model constructor 211 from the three-dimensional model constructor 211. The three-dimensional scene model or the third three-dimensional scene model is subjected to three-dimensional rendering processing to obtain the rendered first three-dimensional scene model or the third three-dimensional scene model.
在一实施方式中,文件存储器201还可以通过上传下载模块202接收来自坐席集成接口105的数据包,并存储在指定目录,其中数据包中包括三维模型信息文件。In one embodiment, the file storage 201 can also receive a data package from the agent integration interface 105 through the upload and download module 202, and store it in a designated directory, where the data package includes a three-dimensional model information file.
在一实施方式中,服务器端200还包括上传下载模块202,上传下载模可以用于根据文件存储器201存储的文件的目录,检测目录层级数,检测服务器端200的目录是否存在,以及检测文件的名称或者格式是否正确,以及令牌的合法性。确认无误后,生成URL(Uniform Resource Locator,统一资源定位符)地址,并把该URL地址发送至设计端100和客户端300。In one embodiment, the server 200 also includes an upload and download module 202. The upload and download module can be used to detect the number of directory levels according to the directory of files stored in the file storage 201, detect whether the directory of the server 200 exists, and detect the file. Whether the name or format is correct, and the token is legal. After confirming that it is correct, generate a URL (Uniform Resource Locator, Uniform Resource Locator) address, and send the URL address to the design end 100 and the client 300.
在一实施方式中,消息处理模块203可以转发账户登录请求;可以分发客户端300发起的查询操作给其他模块,给其他模块发送消息;同时当处理的请求消息过多时,可以把一部 分用户信息及坐席地址等数据存入消息缓存模块204。In one embodiment, the message processing module 203 can forward the account login request; can distribute the query operation initiated by the client 300 to other modules, and send messages to other modules; at the same time, when there are too many request messages to be processed, one can Data such as user information and agent address are stored in the message cache module 204.
在一实施方式中,图像解析模块209还可以应用图像处理方面的算法以及计算机视觉方面的算法,并且该图像解析模型还可以用于图像识别和图像分割,比如,一张来自设计端100的带有红色框标注的对象,经过该图像解析模块209处理后,解析出对象的图形信息,比如该对象的子模型和标注的坐标信息,然后通过消息处理模块203和消息缓存模块204通知客户端300在视频流中去捕获该对象。In one embodiment, the image analysis module 209 can also apply image processing algorithms and computer vision algorithms, and the image analysis model can also be used for image recognition and image segmentation. For example, a tape from the design end 100 After being processed by the image parsing module 209, the object marked with a red frame parses out the graphic information of the object, such as the object's sub-model and the marked coordinate information, and then notifies the client 300 through the message processing module 203 and the message caching module 204. Capture the object in the video stream.
在一实施方式中,图像解析模块209可以从文件存储器201中获取多个视频图像,并对多个视频图像进行分割处理,得到实物图像集合和环境图像集合。In one embodiment, the image analysis module 209 can obtain multiple video images from the file storage 201 and perform segmentation processing on the multiple video images to obtain a set of physical images and a set of environmental images.
在一实施方式中,模型存储模块208包括预设的三维模型库,预设的三维模型库包括实物模型库和环境模型库。In one embodiment, the model storage module 208 includes a preset three-dimensional model library, and the preset three-dimensional model library includes a physical model library and an environment model library.
在一实施例中,设计端100与客户端300建立增强现实视频通话后,设计端100的第一音视频处理模块103可以获取来自坐席集成接口105的增强现实视频通话的视频流或者来自数据存储模块104的的增强现实视频通话的视频流,第一音视频处理模块103对增强现实视频通话的视频流进行抽帧处理,得到多个视频图像,接着第一音视频处理模块103将该多个视频图像通过坐席集成接口105发送至服务器端200。In one embodiment, after the design end 100 and the client 300 establish an augmented reality video call, the first audio and video processing module 103 of the design end 100 can obtain the video stream of the augmented reality video call from the agent integration interface 105 or from the data storage. For the video stream of the augmented reality video call in module 104, the first audio and video processing module 103 performs frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images, and then the first audio and video processing module 103 processes the multiple video images. The video image is sent to the server 200 through the agent integration interface 105.
在一实施例中,文件存储器201可以通过上传下载模块202获取设计端100的坐席集成接口105的多个视频图像,也可以通过上传下载模块202获取客户端300的客户端集成接口302的多个视频图像,在此不做具体限制。In an embodiment, the file storage 201 can obtain multiple video images of the agent integration interface 105 of the design terminal 100 through the upload and download module 202, and can also obtain multiple video images of the client integration interface 302 of the client 300 through the upload and download module 202. Video images are not specifically limited here.
在一实施例中,文件存储器201存储来自设计端100的坐席集成接口105的多个视频图像,或者存储来自客户端300的客户端集成接口302的多个视频图像,图像解析模块209可以从文件存储器201中获取多个视频图像,并对多个视频图像进行分割处理,得到实物图像集合和环境图像集合,并且对实物图像集合进行解析处理得到图形信息,以及对环境图像集合进行解析处理得到标签信息,并将该图形信息和该标签信息发送给模型适配模块210,接着模型适配模块210利用图形信息从模型存储模块208中的实物模型库中获取实物图像元素信息,并且利用实物图像元素信息构建实物图像底图,并对实物图像底图进行建模处理,得到实物三维模型,并且模型适配模块210利用标签信息从模型存储模块208中的环境模型库中获取环境图像元素信息,利用环境图像元素信息对环境图像集合进行建模处理,得到环境三维模型,模型适配模块210将实物三维模型和环境三维模型发送至三维模型构造器211,三维模型构造器211利用实物三维模型和环境三维模型生成第一三维场景模型,并将第一三维场景模型发送至第二三维渲染模块206,通过第二三维渲染模块206对第一三维场景模型进行三维渲染处理,得到渲染后的第一三维场景模型,然后第二三维渲染模块206将该渲染后的第一三维场景模型发送给视频流模块,通过视频流模块发送至设计端100的坐席集成接口105;或者,三维模型构造器211通过上传下载模块202将第一三维场景模型发送至文件存储器201,文件存储器201对第一三维场景模型进行存储,也可以将该第一三维场景模块发送至坐席集成接口105,也可以将该第一三维场景模块发送至客户端300坐席接口。In one embodiment, the file storage 201 stores multiple video images from the agent integration interface 105 of the design end 100, or stores multiple video images from the client integration interface 302 of the client 300. The image parsing module 209 can extract data from the file. Acquire multiple video images in the memory 201, segment the multiple video images to obtain a set of physical images and a set of environmental images, perform analysis on the set of physical images to obtain graphic information, and perform analysis on the set of environmental images to obtain labels. information, and sends the graphic information and the label information to the model adaptation module 210. Then the model adaptation module 210 uses the graphic information to obtain the physical image element information from the physical model library in the model storage module 208, and uses the physical image element The information is used to construct a base map of the physical image, and modeling is performed on the base map of the physical object to obtain a three-dimensional model of the physical object, and the model adaptation module 210 uses the label information to obtain the environmental image element information from the environment model library in the model storage module 208, using The environmental image element information is used to model the environmental image collection to obtain an environmental three-dimensional model. The model adaptation module 210 sends the physical three-dimensional model and the environmental three-dimensional model to the three-dimensional model constructor 211. The three-dimensional model constructor 211 uses the physical three-dimensional model and the environment. The three-dimensional model generates a first three-dimensional scene model, and sends the first three-dimensional scene model to the second three-dimensional rendering module 206. The second three-dimensional rendering module 206 performs three-dimensional rendering processing on the first three-dimensional scene model to obtain the rendered first three-dimensional scene model. scene model, and then the second three-dimensional rendering module 206 sends the rendered first three-dimensional scene model to the video stream module, and sends it to the agent integration interface 105 of the design end 100 through the video stream module; or, the three-dimensional model constructor 211 uploads The download module 202 sends the first three-dimensional scene model to the file storage 201, and the file storage 201 stores the first three-dimensional scene model. The first three-dimensional scene module can also be sent to the agent integration interface 105, or the first three-dimensional scene module can be sent to the agent integration interface 105. The scene module is sent to the client 300 agent interface.
在一实施例中,设计端100的坐席集成接口105可以从视频流模块获取渲染后的第一三维场景模块,并发送给数据存储模块104,第一音视频处理模块103可以从数据存储模块104获取渲染后的第一三维场景模块,第一音视频处理模块103将该渲染后的第一三维场景模块直接发送给第一显示模块101,通过第一显示模块101对渲染后的第一三维场景模型进行显示,在此不做具体限制。In one embodiment, the agent integration interface 105 of the design end 100 can obtain the rendered first three-dimensional scene module from the video stream module and send it to the data storage module 104. The first audio and video processing module 103 can obtain the rendered first three-dimensional scene module from the data storage module 104. Obtaining the rendered first three-dimensional scene module, the first audio and video processing module 103 directly sends the rendered first three-dimensional scene module to the first display module 101, and the first display module 101 displays the rendered first three-dimensional scene module. The model is displayed and no specific restrictions are made here.
在另一实施例中,设计端100的坐席集成接口105可以从数据存储模块104获取渲染后的第一三维场景模块,第一音视频处理模块103可以从数据存储模块104获取渲染后的第一三维场景模块,发送至第一三维渲染模块102,第一三维渲染模块102将渲染后的第一三维场景模型再次进行三维渲染处理,最后第一显示模块101接收来自第一三维渲染模块102的再次渲染后的第一三维场景模型,对该再次渲染后的第一三维场景模型进行显示,在此不做具体限制。 In another embodiment, the agent integration interface 105 of the design end 100 can obtain the rendered first three-dimensional scene module from the data storage module 104, and the first audio and video processing module 103 can obtain the rendered first three-dimensional scene module from the data storage module 104. The three-dimensional scene module is sent to the first three-dimensional rendering module 102. The first three-dimensional rendering module 102 performs three-dimensional rendering processing on the rendered first three-dimensional scene model again. Finally, the first display module 101 receives another three-dimensional rendering process from the first three-dimensional rendering module 102. The rendered first three-dimensional scene model displays the re-rendered first three-dimensional scene model, which is not specifically limited here.
在一实施例中,设计端100的坐席集成接口105可以通过服务器端200的上传下载模块202从服务器端200的文件存储器201获取第一三维场景模块,并发送给第一音视频处理模块103,第一音视频处理模块103将该第一三维场景模块发送给第一三维渲染模块102,通过第一三维渲染模块102对该第一三维场景模型进行三维渲染处理,最后第一显示模块101接收来自第一三维渲染模块102的渲染后的第一三维场景模型,对该渲染后的第一三维场景模型进行显示,在此不做具体限制。In one embodiment, the agent integration interface 105 of the design end 100 can obtain the first three-dimensional scene module from the file storage 201 of the server 200 through the upload and download module 202 of the server 200, and send it to the first audio and video processing module 103. The first audio and video processing module 103 sends the first three-dimensional scene module to the first three-dimensional rendering module 102, and performs three-dimensional rendering processing on the first three-dimensional scene model through the first three-dimensional rendering module 102. Finally, the first display module 101 receives from The rendered first three-dimensional scene model of the first three-dimensional rendering module 102 displays the rendered first three-dimensional scene model, which is not specifically limited here.
在另一实施例中,设计端100的坐席集成接口105可以通过服务器端200的上传下载模块202从服务器端200的文件存储器201获取第一三维场景模块,并发送给数据存储模块104,第一音视频处理模块103从数据存储模块104获取第一三维场景模块,将该第一三维场景模块发送给第一三维渲染模块102,通过第一三维渲染模块102对该第一三维场景模型进行三维渲染处理,最后第一显示模块101接收来自第一三维渲染模块102的渲染后的第一三维场景模型,对该渲染后的第一三维场景模型进行显示,在此不做具体限制。In another embodiment, the agent integration interface 105 of the design end 100 can obtain the first three-dimensional scene module from the file storage 201 of the server 200 through the upload and download module 202 of the server 200, and send it to the data storage module 104. The first The audio and video processing module 103 obtains the first three-dimensional scene module from the data storage module 104, sends the first three-dimensional scene module to the first three-dimensional rendering module 102, and performs three-dimensional rendering on the first three-dimensional scene model through the first three-dimensional rendering module 102. Processing, finally the first display module 101 receives the rendered first three-dimensional scene model from the first three-dimensional rendering module 102, and displays the rendered first three-dimensional scene model, which is not specifically limited here.
在一实施方式中,客户端300可以通过Web软件或者直接在本地终端的3D程序中点击云渲染按钮并借助高速互联网接入访问资源,指令从用户终端中发出,服务器根据指令执行对应的渲染任务,而渲染结果画面则被传送回用户终端中加以显示。把远端的渲染能力提供给终端设备,可以弥补终端的渲染能力短板。In one embodiment, the client 300 can click the cloud rendering button in the 3D program of the local terminal through Web software or directly access the resources through high-speed Internet access. The instructions are issued from the user terminal, and the server executes the corresponding rendering tasks according to the instructions. , and the rendering result screen is sent back to the user terminal for display. Providing remote rendering capabilities to terminal devices can make up for the terminal's shortcomings in rendering capabilities.
在一实施方式中,服务器端200还包括信息库207,其中,信息库207是服务器端200重要的信息数据的存储数据库,为了防止系统异常导致消息缓存模块204中用户信息以及坐席地址等数据丢失,消息缓存模块204可以将一些重要的用户信息以及坐席地址等数据存入信息库207,而且信息库207可以用来恢复中断的业务。In one embodiment, the server 200 also includes an information database 207, where the information database 207 is a storage database for important information data of the server 200. In order to prevent system abnormalities from causing the loss of user information, agent addresses and other data in the message cache module 204 , the message cache module 204 can store some important user information, agent address and other data into the information database 207, and the information database 207 can be used to resume interrupted services.
客户端300包括客户端集成接口302、摄像采集模块303、第二显示模块306、第三三维渲染模块305和第二音视频处理模块304。客户端集成接口302可以用于获取第二音视频处理模块304中的多个视频图像,并对该多个视频图像发送至服务器端200;第二显示模块306可以用于显示渲染后的第一三维场景模型或者第三三维场景模型,其中,显示模块包括电脑显示屏或AR屏幕等。第三三维渲染模块305,用于将来自服务器端200的第一三维场景模型或者第三三维场景模型进行三维渲染处理,得到渲染后的第一三维场景模型或者第三三维场景模型;第二音视频处理模块304可以用于接收对客户端集成接口302发送的增强现实视频通话的视频流,并对增强现实视频通话的视频流进行抽帧处理,得到多个视频图像;摄像采集模块303用于采集增强现实视频通话的视频流。The client 300 includes a client integrated interface 302, a camera acquisition module 303, a second display module 306, a third three-dimensional rendering module 305 and a second audio and video processing module 304. The client integration interface 302 can be used to obtain multiple video images in the second audio and video processing module 304, and send the multiple video images to the server 200; the second display module 306 can be used to display the rendered first A three-dimensional scene model or a third three-dimensional scene model, in which the display module includes a computer display screen or an AR screen. The third three-dimensional rendering module 305 is used to perform three-dimensional rendering processing on the first three-dimensional scene model or the third three-dimensional scene model from the server 200 to obtain the rendered first three-dimensional scene model or the third three-dimensional scene model; the second sound The video processing module 304 can be used to receive the video stream of the augmented reality video call sent to the client integrated interface 302, and perform frame extraction processing on the video stream of the augmented reality video call to obtain multiple video images; the camera acquisition module 303 is used to Capture the video stream of an augmented reality video call.
在一实施方式中,客户端300还包括本地存储模块301,该本地存储模块301用于当从服务器端200文件存储器201中上传或下载文件,根据上传下载模块202发送的URL地址,将服务器端200的文件保存在本地地址或者上传到服务器端200,其中,还可以对文件的格式进行转换,如图片转换为二进制文件。In one embodiment, the client 300 also includes a local storage module 301. The local storage module 301 is used to upload or download files from the file storage 201 of the server 200, according to the URL address sent by the upload and download module 202, to the server. 200 files are saved in the local address or uploaded to the server 200. The format of the files can also be converted, such as converting pictures into binary files.
在一实施方式中,客户端集成接口302可用于接收服务器端200的消息处理模块203的访问请求,并对该访问请求进行解析,还可以获取第一音视频处理模块103中的相关信息进行打包得到数据包,将该数据包发送至服务器端200,在此不做具体限制。In one embodiment, the client integration interface 302 can be used to receive an access request from the message processing module 203 of the server 200, parse the access request, and also obtain relevant information in the first audio and video processing module 103 for packaging. Obtain the data packet and send the data packet to the server 200, which is not specifically limited here.
在一实施例中,客户端300的本地存储模块301可以通过服务器端200的上传下载模块202从服务器端200的文件存储器201下载第一三维场景模块。In an embodiment, the local storage module 301 of the client 300 can download the first three-dimensional scene module from the file storage 201 of the server 200 through the upload and download module 202 of the server 200.
在一实施例中,摄像采集模块303采集增强现实视频通话的视频流,并将该视频流发送至第二音视频处理模块304,第二音视频处理模块304对增强现实视频通话的视频流进行抽帧处理,得到多个视频图像,并多个视频图像发送给客户端集成接口302,客户端集成接口302将该多个视频图像发送至服务器端200,比如,客户端集成接口302将该多个视频图像存储至本地存储模块301,并通过本地存储模块301将该多个视频图像通过服务器端200的上传下载模块202发送至服务器端200的文件存储器201。In one embodiment, the camera collection module 303 collects the video stream of the augmented reality video call, and sends the video stream to the second audio and video processing module 304. The second audio and video processing module 304 processes the video stream of the augmented reality video call. Frame extraction processing is performed to obtain multiple video images, and the multiple video images are sent to the client integration interface 302. The client integration interface 302 sends the multiple video images to the server 200. For example, the client integration interface 302 sends the multiple video images to the server 200. The multiple video images are stored in the local storage module 301, and the multiple video images are sent to the file storage 201 of the server 200 through the upload and download module 202 of the server 200 through the local storage module 301.
在一实施例中,客户端300的客户端集成接口302可以从视频流模块获取渲染后的第一三维场景模块,并发送给第二音视频处理模块304,第二音视频处理模块304将该渲染后的 第一三维场景模块直接发送给第二显示模块306,通过第二显示模块306对渲染后的第一三维场景模型进行显示,在此不做具体限制。In one embodiment, the client integration interface 302 of the client 300 can obtain the rendered first three-dimensional scene module from the video stream module and send it to the second audio and video processing module 304. The second audio and video processing module 304 will Rendered The first three-dimensional scene module is directly sent to the second display module 306, and the second display module 306 displays the rendered first three-dimensional scene model, which is not specifically limited here.
在另一实施例中,客户端300的客户端集成接口302可以从视频流模块获取渲染后的第一三维场景模块,并发送给第二音视频处理模块304,第二音视频处理模块304将该渲染后的第一三维场景模块发送给第三三维渲染模块305,通过第三三维渲染模块305对渲染后的第一三维场景模型再次进行三维渲染处理,最后第二显示模块306接收来自第三三维渲染模块305的再次渲染后的第一三维场景模型,对该再次渲染后的第一三维场景模型进行显示,在此不做具体限制。In another embodiment, the client integration interface 302 of the client 300 can obtain the rendered first three-dimensional scene module from the video stream module and send it to the second audio and video processing module 304. The second audio and video processing module 304 will The rendered first three-dimensional scene module is sent to the third three-dimensional rendering module 305, and the third three-dimensional rendering module 305 performs three-dimensional rendering processing on the rendered first three-dimensional scene model again. Finally, the second display module 306 receives data from the third three-dimensional rendering module 305. The re-rendered first three-dimensional scene model of the three-dimensional rendering module 305 displays the re-rendered first three-dimensional scene model, which is not specifically limited here.
在一实施例中,客户端300的客户端集成接口302可以从本地存储模块301获取第一三维场景模块,并发送给第二音视频处理模块304,第二音视频处理模块304将该第一三维场景模块发送给第三三维渲染模块305,通过第三三维渲染模块305对该第一三维场景模型进行三维渲染处理,最后第二显示模块306接收来自第三三维渲染模块305的渲染后的第一三维场景模型,对该渲染后的第一三维场景模型进行显示,在此不做具体限制。In an embodiment, the client integration interface 302 of the client 300 can obtain the first three-dimensional scene module from the local storage module 301 and send it to the second audio and video processing module 304, and the second audio and video processing module 304 will process the first three-dimensional scene module. The three-dimensional scene module is sent to the third three-dimensional rendering module 305, and the third three-dimensional rendering module 305 performs three-dimensional rendering processing on the first three-dimensional scene model. Finally, the second display module 306 receives the rendered third image from the third three-dimensional rendering module 305. A three-dimensional scene model is used to display the rendered first three-dimensional scene model, which is not specifically limited here.
值得注意的是,在本申请的各个具体实施方式中,当涉及到需要根据用户信息、用户行为数据,用户历史数据以及用户位置信息等与用户身份或特性相关的数据进行相关处理时,都会先获得用户的许可或者同意,而且,对这些数据的收集、使用和处理等,都会遵守相关国家和地区的相关法律法规和标准。此外,当本申请实施例需要获取用户的敏感个人信息时,会通过弹窗或者跳转到确认页面等方式获得用户的单独许可或者单独同意,在明确获得用户的单独许可或者单独同意之后,再获取用于使本申请实施例能够正常运行的必要的用户相关数据。It is worth noting that in each specific implementation of the present application, when it comes to relevant processing based on user information, user behavior data, user historical data, user location information and other data related to user identity or characteristics, the first step is to perform relevant processing. The user's permission or consent is obtained, and the collection, use and processing of this data will comply with the relevant laws, regulations and standards of the relevant countries and regions. In addition, when the embodiment of this application needs to obtain the user's sensitive personal information, it will obtain the user's separate permission or separate consent through a pop-up window or jump to a confirmation page. After clearly obtaining the user's separate permission or separate consent, it will then Obtain necessary user-related data for normal operation of the embodiment of the present application.
另外,参照图11,本申请的一个实施例还提供了另一种三维模型传输装置,该三维模型传输装置400包括存储器402、处理器401及存储在存储器402上并可在处理器401上运行的计算机程序。In addition, referring to Figure 11, one embodiment of the present application also provides another three-dimensional model transmission device. The three-dimensional model transmission device 400 includes a memory 402, a processor 401, and is stored in the memory 402 and can be run on the processor 401. computer program.
处理器401和存储器402可以通过总线或者其他方式连接。The processor 401 and the memory 402 may be connected through a bus or other means.
存储器402作为一种非暂态计算机可读存储介质,可用于存储非暂态软件程序以及非暂态性计算机可执行程序。此外,存储器402可以包括高速随机存取存储器,还可以包括非暂态存储器,例如至少一个磁盘存储器件、闪存器件、或其他非暂态固态存储器件。在一些实施方式中,存储器402可包括相对于处理器401远程设置的存储器,这些远程存储器可以通过网络连接至该处理器401。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。As a non-transitory computer-readable storage medium, the memory 402 can be used to store non-transitory software programs and non-transitory computer executable programs. In addition, memory 402 may include high-speed random access memory and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid-state storage device. In some embodiments, the memory 402 may include memory located remotely relative to the processor 401, and these remote memories may be connected to the processor 401 through a network. Examples of the above-mentioned networks include but are not limited to the Internet, intranets, local area networks, mobile communication networks and combinations thereof.
实现上述实施例的三维模型传输方法所需的非暂态软件程序以及指令存储在存储器402中,当被处理器401执行时,执行上述实施例中的三维模型传输方法,例如,执行以上描述的图1中的方法步骤S110至S160、图2中的方法步骤S210至S230、图3中的方法步骤S310至S350、图4中的方法步骤S410至S440、图5中的方法步骤S510至S530、图6中的方法步骤S610至S620、图7中的方法步骤S710至S720以及图8中的方法步骤S810至S820。The non-transient software programs and instructions required to implement the three-dimensional model transmission method in the above embodiment are stored in the memory 402. When executed by the processor 401, the three-dimensional model transmission method in the above embodiment is executed, for example, the above-described The method steps S110 to S160 in Figure 1, the method steps S210 to S230 in Figure 2, the method steps S310 to S350 in Figure 3, the method steps S410 to S440 in Figure 4, the method steps S510 to S530 in Figure 5, Method steps S610 to S620 in FIG. 6 , method steps S710 to S720 in FIG. 7 , and method steps S810 to S820 in FIG. 8 .
以上所描述的设备实施例仅仅是示意性的,其中作为分离部件说明的单元可以是或者也可以不是物理上分开的,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部模块来实现本实施例方案的目的。The device embodiments described above are only illustrative, and the units described as separate components may or may not be physically separate, that is, they may be located in one place, or they may be distributed to multiple network units. Some or all of the modules can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
此外,本申请的一个实施例还提供了一种计算机可读存储介质,该计算机可读存储介质存储有计算机可执行指令,该计算机可执行指令被一个处理器或控制器执行,例如,被上述设备实施例中的一个处理器执行,可使得上述处理器执行上述实施例中的三维模型传输方法,执行以上描述的图1中的方法步骤S110至S160、图2中的方法步骤S210至S230、图3中的方法步骤S310至S350、图4中的方法步骤S410至S440、图5中的方法步骤S510至S530、图6中的方法步骤S610至S620、图7中的方法步骤S710至S720以及图8中的方法步骤S810至S820。In addition, an embodiment of the present application also provides a computer-readable storage medium that stores computer-executable instructions, and the computer-executable instructions are executed by a processor or controller, for example, by the above-mentioned Execution by a processor in the device embodiment can cause the above-mentioned processor to execute the three-dimensional model transmission method in the above embodiment, and execute the above-described method steps S110 to S160 in Figure 1, method steps S210 to S230 in Figure 2, The method steps S310 to S350 in Figure 3, the method steps S410 to S440 in Figure 4, the method steps S510 to S530 in Figure 5, the method steps S610 to S620 in Figure 6, the method steps S710 to S720 in Figure 7 and Method steps S810 to S820 in Figure 8 .
此外,本申请的一个实施例还提供了一种计算机程序产品,包括计算机程序或计算机指 令,计算机程序或计算机指令存储在计算机可读存储介质中,计算机设备的处理器从计算机可读存储介质读取计算机程序或计算机指令,处理器执行计算机程序或计算机指令,使得计算机设备执行上述实施例中的三维模型传输方法,例如,执行以上描述的图1中的方法步骤S110至S160、图2中的方法步骤S210至S230、图3中的方法步骤S310至S350、图4中的方法步骤S410至S440、图5中的方法步骤S510至S530、图6中的方法步骤S610至S620、图7中的方法步骤S710至S720以及图8中的方法步骤S810至S820。In addition, an embodiment of the present application also provides a computer program product, including a computer program or computer instructions Let the computer program or computer instructions be stored in a computer-readable storage medium, the processor of the computer device reads the computer program or computer instructions from the computer-readable storage medium, and the processor executes the computer program or computer instructions, causing the computer device to perform the above implementation The three-dimensional model transmission method in the example, for example, executes the above-described method steps S110 to S160 in Figure 1, method steps S210 to S230 in Figure 2, method steps S310 to S350 in Figure 3, and method steps in Figure 4. S410 to S440, method steps S510 to S530 in FIG. 5 , method steps S610 to S620 in FIG. 6 , method steps S710 to S720 in FIG. 7 , and method steps S810 to S820 in FIG. 8 .
本申请实施例包括:首先服务器端获取来自设计端的多个视频图像,视频图像由设计端通过对增强现实视频通话的视频流进行抽帧处理而获得,增强现实视频通话由设计端与客户端建立,接着对多个视频图像进行分割处理,得到实物图像集合和环境图像集合,然后对实物图像集合进行建模处理,得到实物三维模型,并且对环境图像集合进行建模处理,得到环境三维模型,根据实物三维模型和环境三维模型生成第一三维场景模型,最后将第一三维场景模型发送至客户端,即是说,在设计端与客户端建立的增强现实视频通话的场景下,设计端对该增强现实视频通话的视频流进行抽帧处理得到多个视频图像,服务器端通过对该多个视频图像进行处理,最终得到第一三维场景模型,并将该第一三维场景模型发送至客户端,因此,本申请实施例能够在增强现实场景下实现三维模型的传输。The embodiments of this application include: first, the server side obtains multiple video images from the design side. The video images are obtained by the design side by performing frame extraction processing on the video stream of the augmented reality video call. The augmented reality video call is established by the design side and the client. , then perform segmentation processing on multiple video images to obtain a collection of physical images and a collection of environmental images, then perform modeling processing on the collection of physical images to obtain a three-dimensional model of the physical object, and perform modeling processing on the collection of environmental images to obtain a three-dimensional model of the environment, The first 3D scene model is generated based on the physical 3D model and the environment 3D model, and finally the first 3D scene model is sent to the client. That is to say, in the scenario of the augmented reality video call established between the design end and the client, the design end The video stream of the augmented reality video call is subjected to frame extraction processing to obtain multiple video images. The server processes the multiple video images to finally obtain a first three-dimensional scene model, and sends the first three-dimensional scene model to the client. , Therefore, the embodiments of the present application can realize the transmission of three-dimensional models in augmented reality scenarios.
本领域普通技术人员可以理解,上文中所公开方法中的全部或某些步骤、系统可以被实施为软件、固件、硬件及其适当的组合。某些物理组件或所有物理组件可以被实施为由处理器,如中央处理器、数字信号处理器或微处理器执行的软件,或者被实施为硬件,或者被实施为集成电路,如专用集成电路。这样的软件可以分布在计算机可读介质上,计算机可读介质可以包括计算机存储介质(或非暂时性介质)和通信介质(或暂时性介质)。如本领域普通技术人员公知的,术语计算机存储介质包括在用于存储信息(诸如计算机可读指令、数据结构、程序模块或其他数据)的任何方法或技术中实施的易失性和非易失性、可移除和不可移除介质。计算机存储介质包括但不限于RAM、ROM、EEPROM、闪存或其他存储器技术、CD-ROM、数字多功能盘(DVD)或其他光盘存储、磁盒、磁带、磁盘存储或其他磁存储装置、或者可以用于存储期望的信息并且可以被计算机访问的任何其他的介质。此外,本领域普通技术人员公知的是,通信介质通常包含计算机可读指令、数据结构、程序模块或者诸如载波或其他传输机制之类的调制数据信号中的其他数据,并且可包括任何信息递送介质。 Those of ordinary skill in the art can understand that all or some steps and systems in the methods disclosed above can be implemented as software, firmware, hardware, and appropriate combinations thereof. Some or all of the physical components may be implemented as software executed by a processor, such as a central processing unit, a digital signal processor, or a microprocessor, or as hardware, or as an integrated circuit, such as an application specific integrated circuit . Such software may be distributed on computer-readable media, which may include computer storage media (or non-transitory media) and communication media (or transitory media). As is known to those of ordinary skill in the art, the term computer storage media includes volatile and nonvolatile media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data. removable, removable and non-removable media. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, Digital Versatile Disk (DVD) or other optical disk storage, magnetic cassettes, tapes, disk storage or other magnetic storage devices, or may Any other medium used to store the desired information and that can be accessed by a computer. Additionally, it is known to those of ordinary skill in the art that communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism, and may include any information delivery media .

Claims (12)

  1. 一种三维模型传输方法,包括:A three-dimensional model transmission method, including:
    获取来自设计端的多个视频图像,所述视频图像由所述设计端通过对增强现实视频通话的视频流进行抽帧处理而获得,所述增强现实视频通话由所述设计端与客户端建立;Acquire multiple video images from the design end, the video images are obtained by the design end by performing frame extraction processing on the video stream of the augmented reality video call, and the augmented reality video call is established by the design end and the client;
    对多个所述视频图像进行分割处理,得到实物图像集合和环境图像集合;Perform segmentation processing on multiple video images to obtain a collection of physical images and a collection of environmental images;
    对所述实物图像集合进行建模处理,得到实物三维模型;Perform modeling processing on the collection of real object images to obtain a three-dimensional model of the real object;
    对所述环境图像集合进行建模处理,得到环境三维模型;Perform modeling processing on the environmental image collection to obtain a three-dimensional environment model;
    根据所述实物三维模型和所述环境三维模型生成第一三维场景模型;Generate a first three-dimensional scene model based on the physical three-dimensional model and the environmental three-dimensional model;
    将所述第一三维场景模型发送至所述客户端。Send the first three-dimensional scene model to the client.
  2. 根据权利要求1所述的三维模型传输方法,其中,所述根据所述实物三维模型和所述环境三维模型生成第一三维场景模型,包括:The three-dimensional model transmission method according to claim 1, wherein generating the first three-dimensional scene model according to the physical three-dimensional model and the environmental three-dimensional model includes:
    获取预设的三维模型库,所述三维模型库包括环境模型库;Obtain a preset three-dimensional model library, where the three-dimensional model library includes an environment model library;
    根据所述实物三维模型和所述环境三维模型从所述环境模型库中确定目标环境三维模型;Determine a target environment three-dimensional model from the environment model library according to the physical three-dimensional model and the environment three-dimensional model;
    根据所述实物三维模型和所述目标环境三维模型生成第一三维场景模型。A first three-dimensional scene model is generated according to the three-dimensional model of the physical object and the three-dimensional model of the target environment.
  3. 根据权利要求1所述的三维模型传输方法,其中,在所述将所述第一三维场景模型发送至所述客户端之后,所述三维模型传输方法还包括:The three-dimensional model transmission method according to claim 1, wherein after sending the first three-dimensional scene model to the client, the three-dimensional model transmission method further includes:
    接收所述客户端发送的标注信息;Receive annotation information sent by the client;
    根据所述标注信息在所述第一三维场景模型中确定标注区域;Determine a labeling area in the first three-dimensional scene model according to the labeling information;
    在所述标注区域上嵌套所述标注信息,得到第二三维场景模型;Nest the annotation information on the annotation area to obtain a second three-dimensional scene model;
    将所述第一三维场景模型和所述第二三维场景模型进行叠加处理,得到第三三维场景模型;Superimpose the first three-dimensional scene model and the second three-dimensional scene model to obtain a third three-dimensional scene model;
    将所述第三三维场景模型发送至所述客户端。Send the third three-dimensional scene model to the client.
  4. 根据权利要求2所述的三维模型传输方法,其中,所述三维模型库包括实物模型库,所述对所述实物图像集合进行建模处理,得到实物三维模型,包括:The three-dimensional model transmission method according to claim 2, wherein the three-dimensional model library includes a physical model library, and the modeling process on the physical image collection to obtain the physical three-dimensional model includes:
    对所述实物图像集合进行解析处理,得到图形信息;Analyze and process the collection of physical images to obtain graphic information;
    根据所述图形信息从所述实物模型库中获取实物图像元素信息;Obtain physical image element information from the physical model library according to the graphic information;
    利用所述实物图像元素信息构建实物图像底图;Using the physical image element information to construct a physical image base map;
    对所述实物图像底图进行建模处理,得到实物三维模型。Perform modeling processing on the image base map of the physical object to obtain a three-dimensional model of the physical object.
  5. 根据权利要求2所述的三维模型传输方法,其中,所述对所述环境图像集合进行建模处理,得到环境三维模型,包括:The three-dimensional model transmission method according to claim 2, wherein the modeling process on the environmental image collection to obtain the environmental three-dimensional model includes:
    对所述环境图像集合进行解析处理,得到标签信息;Perform analysis and processing on the environmental image collection to obtain label information;
    根据所述标签信息从所述环境模型库中获取环境图像元素信息;Obtain environmental image element information from the environment model library according to the tag information;
    利用所述环境图像元素信息对所述环境图像集合进行建模处理,得到环境三维模型。The environmental image element information is used to perform modeling processing on the environmental image set to obtain an environmental three-dimensional model.
  6. 根据权利要求2所述的三维模型传输方法,其中,所述三维模型传输方法还包括:The three-dimensional model transmission method according to claim 2, wherein the three-dimensional model transmission method further includes:
    根据所述实物三维模型从所述环境模型库中确定目标环境三维模型;Determine a target environment three-dimensional model from the environment model library according to the physical three-dimensional model;
    将所述实物三维模型与所述目标环境三维模型进行嵌套处理,生成所述第一三维场景模型。The physical three-dimensional model and the target environment three-dimensional model are nested to generate the first three-dimensional scene model.
  7. 根据权利要求1所述的三维模型传输方法,其中,所述根据所述实物三维模型和所述环境三维模型生成第一三维场景模型,包括:The three-dimensional model transmission method according to claim 1, wherein generating the first three-dimensional scene model according to the physical three-dimensional model and the environmental three-dimensional model includes:
    将所述实物三维模型与所述环境三维模型进行嵌套处理,生成所述第一三维场景模型。The physical three-dimensional model and the environmental three-dimensional model are nested to generate the first three-dimensional scene model.
  8. 一种三维模型传输方法,包括:A three-dimensional model transmission method, including:
    与设计端建立增强现实视频通话;Establish an augmented reality video call with the designer;
    接收服务器端发送的第一三维场景模型,所述第一三维场景模型由所述服务器端根据实物三维模型和环境三维模型而获得,所述实物三维模型由所述服务器端通过对实物图像集合 进行建模处理而获得,所述环境三维模型由所述服务器端通过对环境图像集合进行建模处理而获得,所述实物图像集合和所述环境图像集合均由所述服务器端通过对多个视频图像进行分割处理而获得,所述视频图像由所述设计端通过对所述增强现实视频通话的视频流进行抽帧处理而获得。Receive the first three-dimensional scene model sent by the server. The first three-dimensional scene model is obtained by the server based on the physical three-dimensional model and the environmental three-dimensional model. The physical three-dimensional model is obtained by the server through a collection of physical images. The environment three-dimensional model is obtained by performing modeling processing on the environment image set by the server side. Both the physical image set and the environment image set are obtained by the server side by modeling a plurality of environment image sets. The video image is obtained by segmentation processing, and the video image is obtained by the design end by performing frame extraction processing on the video stream of the augmented reality video call.
  9. 根据权利要求8所述的三维模型传输方法,其中,所述三维模型传输方法还包括:The three-dimensional model transmission method according to claim 8, wherein the three-dimensional model transmission method further includes:
    对所述第一三维场景模型进行虚拟标注处理,得到标注信息;Perform virtual annotation processing on the first three-dimensional scene model to obtain annotation information;
    将所述标注信息发送至所述服务器端。Send the annotation information to the server.
  10. 一种三维模型传输装置,包括:存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现如权利要求1至9中任意一项所述的三维模型传输方法。A three-dimensional model transmission device, including: a memory, a processor, and a computer program stored in the memory and executable on the processor. When the processor executes the computer program, it implements any one of claims 1 to 9 The three-dimensional model transmission method.
  11. 一种计算机可读存储介质,存储有计算机可执行指令,所述计算机可执行指令用于执行权利要求1至9中任意一项所述的三维模型传输方法。A computer-readable storage medium stores computer-executable instructions, and the computer-executable instructions are used to execute the three-dimensional model transmission method described in any one of claims 1 to 9.
  12. 一种计算机程序产品,包括计算机程序或计算机指令,所述计算机程序或所述计算机指令存储在计算机可读存储介质中,计算机设备的处理器从所述计算机可读存储介质读取所述计算机程序或所述计算机指令,所述处理器执行所述计算机程序或所述计算机指令,使得所述计算机设备执行如权利要求1至9任意一项所述的三维模型传输方法。 A computer program product comprising a computer program or computer instructions stored in a computer-readable storage medium from which a processor of a computer device reads the computer program Or the computer instruction, the processor executes the computer program or the computer instruction, so that the computer device executes the three-dimensional model transmission method according to any one of claims 1 to 9.
PCT/CN2023/097180 2022-06-08 2023-05-30 Three-dimensional model transmission method and apparatus, and storage medium and program product WO2023236815A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210640757.1A CN117240831A (en) 2022-06-08 2022-06-08 Three-dimensional model transmission method and device, storage medium and program product thereof
CN202210640757.1 2022-06-08

Publications (1)

Publication Number Publication Date
WO2023236815A1 true WO2023236815A1 (en) 2023-12-14

Family

ID=89083145

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/097180 WO2023236815A1 (en) 2022-06-08 2023-05-30 Three-dimensional model transmission method and apparatus, and storage medium and program product

Country Status (2)

Country Link
CN (1) CN117240831A (en)
WO (1) WO2023236815A1 (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105933637A (en) * 2016-04-26 2016-09-07 上海与德通讯技术有限公司 Video communication method and system
US20180197341A1 (en) * 2016-06-10 2018-07-12 Dirtt Environmental Solutions, Ltd. Mixed-reality architectural design environment
CN112446939A (en) * 2020-11-19 2021-03-05 深圳市中视典数字科技有限公司 Three-dimensional model dynamic rendering method and device, electronic equipment and storage medium
CN112907751A (en) * 2021-03-23 2021-06-04 中德(珠海)人工智能研究院有限公司 Virtual decoration method, system, equipment and medium based on mixed reality
CN113242398A (en) * 2021-04-16 2021-08-10 杭州易现先进科技有限公司 Three-dimensional labeled audio and video call method and system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105933637A (en) * 2016-04-26 2016-09-07 上海与德通讯技术有限公司 Video communication method and system
US20180197341A1 (en) * 2016-06-10 2018-07-12 Dirtt Environmental Solutions, Ltd. Mixed-reality architectural design environment
CN112446939A (en) * 2020-11-19 2021-03-05 深圳市中视典数字科技有限公司 Three-dimensional model dynamic rendering method and device, electronic equipment and storage medium
CN112907751A (en) * 2021-03-23 2021-06-04 中德(珠海)人工智能研究院有限公司 Virtual decoration method, system, equipment and medium based on mixed reality
CN113242398A (en) * 2021-04-16 2021-08-10 杭州易现先进科技有限公司 Three-dimensional labeled audio and video call method and system

Also Published As

Publication number Publication date
CN117240831A (en) 2023-12-15

Similar Documents

Publication Publication Date Title
CN109327727B (en) Live stream processing method in WebRTC and stream pushing client
US9990760B2 (en) Generating a 3D interactive immersive experience from a 2D static image
WO2018033138A1 (en) Data processing method, apparatus and electronic device
US20210264139A1 (en) Creating videos with facial expressions
US8878897B2 (en) Systems and methods for sharing conversion data
US20130057642A1 (en) Video conferencing system, method, and computer program storage device
WO2017101487A1 (en) Method and device for submitting transcoding attribute information
CN111464828A (en) Virtual special effect display method, device, terminal and storage medium
AU2015298291A1 (en) Optimized rendering of shared documents on client devices with document raster representations
WO2019193364A1 (en) Method and apparatus for generating augmented reality images
WO2022088834A1 (en) Dynamic photograph album generation method, server, display terminal and readable storage medium
KR20120099814A (en) Augmented reality contents service system and apparatus and method
CN113965773A (en) Live broadcast display method and device, storage medium and electronic equipment
CN114139491A (en) Data processing method, device and storage medium
WO2023236815A1 (en) Three-dimensional model transmission method and apparatus, and storage medium and program product
CN114546577B (en) Data visualization method and system
US11615167B2 (en) Media creation system and method
TWI790560B (en) Side by side image detection method and electronic apparatus using the same
CN112651801B (en) Method and device for displaying house source information
TWI765230B (en) Information processing device, information processing method, and information processing program
JP6623905B2 (en) Server device, information processing method and program
US20230195856A1 (en) Method for media creation, sharing, and communication and associated system
CN117437342B (en) Three-dimensional scene rendering method and storage medium
EP4376385A1 (en) System and method enabling live broadcasting sessions in virtual environments
CN117793481A (en) Video stream generation method, device, equipment and readable storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23818980

Country of ref document: EP

Kind code of ref document: A1