CN113254828B - Seamless multi-mode content mixing exhibition method based on nonlinear editing technology - Google Patents

Seamless multi-mode content mixing exhibition method based on nonlinear editing technology Download PDF

Info

Publication number
CN113254828B
CN113254828B CN202110562948.6A CN202110562948A CN113254828B CN 113254828 B CN113254828 B CN 113254828B CN 202110562948 A CN202110562948 A CN 202110562948A CN 113254828 B CN113254828 B CN 113254828B
Authority
CN
China
Prior art keywords
node
modal
nonlinear
nodes
time
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110562948.6A
Other languages
Chinese (zh)
Other versions
CN113254828A (en
Inventor
赵海英
皮帮瀚
梁昊光
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
BEIJING INTERNATIONAL STUDIES UNIVERSITY
Beijing University of Posts and Telecommunications
Original Assignee
BEIJING INTERNATIONAL STUDIES UNIVERSITY
Beijing University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by BEIJING INTERNATIONAL STUDIES UNIVERSITY, Beijing University of Posts and Telecommunications filed Critical BEIJING INTERNATIONAL STUDIES UNIVERSITY
Priority to CN202110562948.6A priority Critical patent/CN113254828B/en
Publication of CN113254828A publication Critical patent/CN113254828A/en
Application granted granted Critical
Publication of CN113254828B publication Critical patent/CN113254828B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/958Organisation or management of web site content, e.g. publishing, maintaining pages or automatic linking

Abstract

A seamless multi-modal content mixing presentation method based on a non-linear editing technology comprises the following steps: firstly, reading, storing and processing multi-mode data by using a multi-mode content presentation method supported by multiple formats; then, by utilizing a nonlinear editing technology, only multi-modal data needs to be uploaded once, so that the single time sequence editing limit can be broken through, multiple times of arrangement and editing are carried out according to various sequences, and the nonlinear structure is constructed and presented in a mixed mode through a corresponding structure; and finally, rendering the multi-modal content to the page efficiently by using a Canvas layering-based rendering method. This method is a method for editing multi-modal content by a nonlinear structure, and has the characteristics of rapidness, simplicity, and high degree of freedom, as compared with linear editing performed in a fixed time sequence. The embodiment of the invention can effectively solve the problems of low efficiency and single form of making the augmented reality content at the webpage end, so that a user can automatically experience the multimode Web AR content with a story line.

Description

Seamless multi-mode content mixing exhibition method based on nonlinear editing technology
Technical Field
The invention relates to the technical field of virtual reality, in particular to a seamless multi-modal content mixed exhibition method based on a nonlinear editing technology, which relates to the application of augmented reality at a webpage end.
Background
Cultural relics, which are mental resources of culture, are subordinate to culture and are components of culture. The digital propagation of cultural heritage is an indispensable part in the nation of self confidence and strong culture. In recent years, the development of virtual Reality related technologies is rapid, and among them, the Augmented Reality (Augmented Reality) technology becomes one of the very good expression ways for the digitization of cultural relics. The augmented reality (Web AR) at the webpage end has the advantages of natural cross-platform and light weight because the user can experience the augmented reality effect only through the webpage without any software. The combination of cultural relic digital exhibition and Web AR technology is undoubtedly one of the best exhibition forms in the cultural relic digital propagation field.
However, the cultural relic display system based on the Web AR also has a plurality of disadvantages: the method has the advantages of high threshold, low development efficiency, high maintenance cost, complex multi-mode data and few sharing means. The existing Web AR cultural relic display system has the problems of complex multi-modal data and difficult effective integration of contents; this not only makes the Web AR content production form single, but also makes it difficult to produce quality content having a story line.
Disclosure of Invention
The invention aims to solve the problems in the Web AR content production process, and provides a seamless multi-modal content mixing presentation method based on a nonlinear editing technology.
The invention provides a seamless multi-modal content mixing exhibition method based on a nonlinear editing technology, which is characterized by comprising the following steps:
step 1, reading, storing and calculating multi-mode data suitable for Web AR, wherein the multi-mode data comprises a three-dimensional model, pictures, characters, videos and audios;
step 2, firstly constructing nonlinear structures of the modal data nodes, then combining the independent modal data nodes according to the constructed nonlinear structures, and finally, mixing and presenting the data in the Web AR scene according to the configuration resource information of each mode, wherein the specific process is as follows:
2.1, constructing a nonlinear structure of multi-modal data, namely firstly defining a node structure of each modal data, defining a universal node structure and refining the attribute of each node according to the characteristic of each modal data, wherein each attribute and the meaning of the universal node structure are respectively as follows:
(1) NodeId: a unique identifier of a node for distinguishing different nodes;
(2) NodeName: the name of the node;
(3) NodeType: the types of the nodes at least comprise 7 types of the nodes: Start-Start node, End-End node, Model-three-dimensional Model node, Picture-Picture node, Text-Text node, Video-Video node and Audio-Audio node;
(4) NodeLife: the life cycle of a node is a four-tuple: nodelength (lifeLength, canSkip, startTime, endTime), where lifeLength is the life length of a node, i.e., the duration; canSkip indicates whether the node can be skipped; startTime and endTime respectively represent the start time and end time of a node;
(5) NodeProps: the attributes of the nodes at least comprise fileName-fileName, fileType-file type, filePath-file storage path, upLoader-upLoader, uplodeTime-uploading time;
(6) NodeConnections: the association attribute with the current node is integrally represented by a set, each element in the set is a binary connection (source, target), the source is a source node, the target is a target node, and the binary represents a directed connection line associated with the current node;
(7) NodeInfo: node information including node creation time and creator information;
2.2, a two-tuple is used to represent that a total time period T of a nonlinear structure is (start, end), where start and end correspond to a start time point and an end time point of the total time period of the nonlinear structure, respectively, and there is start < end, then a total duration L of the nonlinear structure is end-start, and possible timing relationships for durations d1 and d2 of any two nodes n1 and n2 in the same nonlinear structure are as follows:
(1) neighbor (d1, d 2): n1 and n2 are in a sequential order relationship, so that d2 starts immediately after d1 ends;
(2) gradation (d1, d 2): n1 and n2 are in sequential relation but not adjacent, then d2 begins after a period of time after d1 ends;
(3) async (d1, d 2): n1 and n2 have no sequence relation, the d1 and d2 respectively execute and are not influenced by each other;
according to the time sequence relation, for the data of the five modes, firstly, displaying a three-dimensional model, then, sequentially displaying texts and pictures, simultaneously playing audio, and finally, displaying videos after the pictures and the audio are all finished so as to obtain time nodes related to the display time sequence; analyzing the sequence and the duration of the time nodes to obtain the relationship between the duration of each node of the nonlinear structure and the total time period T; finally, obtaining the logic combination flow of each modal node through the conversion of time mapping;
step 3, adopting a Canvas-based layered rendering method, integrating three-dimensional scenes of a WebRTC video stream and a WebGL and webpage DOM elements, and realizing the layered rendering of a Web AR page and the mixed presentation of a nonlinear structure, wherein the Canvas-based layered implementation method comprises the following specific steps:
3.1, creating a < canvas > element and adding it to the DOM structure;
3.2, adding size and positioning attributes to the < canvas > element to support layering;
3.3, adding a transparency style element to the canvas to generate a transparent background;
3.4, repeating the three steps to generate a multilayer canvas;
and 3.5, binding the node data needing to be hierarchically rendered and displayed on different Canvas hierarchies of the hierarchies through a logic combination flow of each modal node of the nonlinear structure.
The present invention is characterized in that it is fast, simple, and highly flexible to perform multi-modal content editing by a nonlinear structure, as compared with linear editing performed in a certain time sequence. Therefore, the seamless multi-mode content mixed exhibition method based on the nonlinear editing technology can effectively solve the problems of low production efficiency and single form of the augmented reality (Web AR) content at the webpage end, so that a user can automatically experience the multi-mode Web AR content with a story main line.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
Fig. 1 is a flowchart of a seamless multimodal content mixing presentation method based on a nonlinear editing technology according to an embodiment of the present invention.
FIG. 2 is a schematic diagram of a time node with timing dependent display.
Fig. 3 is a schematic diagram of the relationship between the duration of each node of the nonlinear structure and the total time period T.
Fig. 4 is a schematic diagram of a logic combination flow of each modal node.
Detailed Description
The technical solutions in the embodiments of the present invention will be described clearly and completely with reference to the accompanying drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Fig. 1 is a flowchart of a seamless multimodal content mixing presentation method based on a nonlinear editing technology according to an embodiment of the present invention. The principle of the method is as follows: firstly, reading, storing and processing multi-mode data by using a multi-mode content presentation method supported by multiple formats; then, constructing a nonlinear structure by utilizing a nonlinear editing technology and performing mixed presentation by a corresponding structure through a time sequence control method; and finally, rendering the multi-modal content to the page efficiently by using a Canvas layering-based rendering method.
Specifically, the seamless multimodal content mixing presentation method based on the nonlinear editing technology in the embodiment includes the following steps:
s110, reading, storing, calculating and processing multi-mode content suitable for Web AR, including 5 types of modal data such as models, pictures, characters, videos and audios; taking the first four hexadecimal codes of the header of the binary file to judge the accurate type of the file for the multi-modal data file to be processed; and for model data, parsing, storing and rendering in three formats, namely obj, fbx, glTF and the like, are provided.
And S120, defining a general node structure of each modal data, and refining the attribute of the node according to the characteristic of each modal data. The method comprises the following specific steps:
the nonlinear structure of the multi-modal data is constructed by firstly defining the node structure of each modal data, defining a universal node structure and refining the attribute of the node according to the characteristic of each modal data, wherein each attribute and the meaning of the universal node structure are respectively as follows:
(1) NodeId: a unique identifier of a node for distinguishing different nodes;
(2) NodeName: the name of the node;
(3) NodeType: the types of the nodes at least comprise 7 types of the nodes: Start-Start node, End-End node, Model-three-dimensional Model node, Picture-Picture node, Text-Text node, Video-Video node and Audio-Audio node;
(4) NodeLife: the life cycle of a node is a four-tuple: nodelength (lifeLength, canSkip, startTime, endTime), where lifeLength is the life length of a node, i.e., the duration; canSkip indicates whether the node can be skipped; startTime and endTime respectively represent the start time and the end time of a node;
(5) NodeProps: the attributes of the nodes at least comprise fileName-fileName, fileType-file type, filePath-file storage path, upLoader-upLoader, uplodeTime-uploading time;
(6) NodeConnections: the association attribute with the current node is integrally represented by a set, each element in the set is a binary connection (source, target), the source is a source node, the target is a target node, and the binary represents a directed connection line associated with the current node;
(7) NodeInfo: node information including node creation time and creator information;
s130, combining the independent modal data nodes according to the constructed nonlinear structure, and displaying in a Web AR scene according to the configuration resource information of each mode in a mixed mode, wherein the specific process is as follows:
using a two-tuple to represent a total time period T of a nonlinear structure (start, end), where start and end correspond to a start time point and an end time point of the total time period of the nonlinear structure, respectively, and if there is start < end, then the total duration L of the nonlinear structure is end-start, and for durations d1 and d2 of any two nodes n1 and n2 in the same nonlinear structure, the possible timing relationship is as follows:
(1) neighbor (d1, d 2): n1 and n2 are in a sequential order relationship, so that d2 starts immediately after d1 ends;
(2) gradation (d1, d 2): n1 and n2 are in sequential relation but not adjacent, then d2 begins after a period of time after d1 ends;
(3) async (d1, d 2): n1 and n2 have no sequential relation, then d1 and d2 respectively execute and are not influenced by each other;
according to the time sequence relationship, for the data of the five modes, firstly, the three-dimensional model is displayed, then, the text and the picture are sequentially displayed, the audio is played, and finally, after the picture and the audio are finished, the video is displayed, so that the time node (shown in fig. 2) related to the display time sequence is obtained.
Analyzing the sequence and the duration of the time nodes to obtain the relationship between the duration of each node of the nonlinear structure and the total time period T (as shown in FIG. 3); and finally, obtaining the logic combination flow of each modal node through the conversion of the time mapping (as shown in fig. 4).
S140, adopting a Canvas-based layered rendering method, integrating three-dimensional scenes of a WebRTC video stream and a WebGL and webpage DOM elements, and realizing the layered rendering of a Web AR page and the mixed presentation of a nonlinear structure, wherein the Canvas-based layered implementation method comprises the following specific steps:
3.1, creating and adding a < canvas > element to the DOM structure;
3.2, adding size and positioning attributes to the < canvas > element to support layering;
3.3, adding a transparency style element to the canvas to generate a transparent background;
3.4, repeating the three steps to generate a multilayer canvas;
and 3.5, binding the node data needing to be subjected to layered rendering display on different Canvas layers of the layers through the logic combination flow of each modal node of the nonlinear structure.
According to the time sequence relation, for the data of the five modes, firstly, three-dimensional model display is carried out, then, text and picture display is carried out sequentially, audio is played, and finally, after the pictures and the audio are finished, video display is carried out;
in addition to the above embodiments, the present invention may have other embodiments. All technical solutions formed by adopting equivalent substitutions or equivalent transformations fall within the protection scope of the claims of the present invention.

Claims (3)

1. A seamless multi-modal content mixing presentation method based on a non-linear editing technology is characterized by comprising the following steps:
step 1, reading, storing and calculating multi-mode data suitable for Web AR, wherein the multi-mode data comprises a three-dimensional model, pictures, characters, videos and audios;
step 2, firstly constructing nonlinear structures of the modal data nodes, then combining the independent modal data nodes according to the constructed nonlinear structures, and finally, mixing and presenting the data in the Web AR scene according to the configuration resource information of each mode, wherein the specific process is as follows:
2.1, constructing a nonlinear structure of multi-modal data, namely firstly defining a node structure of each modal data, defining a universal node structure and refining the attribute of each node according to the characteristic of each modal data, wherein each attribute and the meaning of the universal node structure are respectively as follows:
(1) NodeId: a unique identifier of a node for distinguishing different nodes;
(2) NodeName: the name of the node;
(3) NodeType: the types of the nodes at least comprise 7 types of the nodes: Start-Start node, End-End node, Model-three-dimensional Model node, Picture-Picture node, Text-Text node, Video-Video node and Audio-Audio node;
(4) NodeLife: the life cycle of a node is a four-tuple: nodelength (lifeLength, canSkip, startTime, endTime), where lifeLength is the life length of a node, i.e., the duration; canSkip indicates whether the node can be skipped; startTime and endTime respectively represent the start time and end time of a node;
(5) NodeProps: the attributes of the nodes at least comprise fileName-fileName, fileType-file type, filePath-file storage path, upLoader-upLoader, uplodeTime-uploading time;
(6) node connections: the association attribute with the current node is integrally represented by a set, each element in the set is a binary connection (source, target), the source is a source node, the target is a target node, and the binary represents a directed connection line associated with the current node;
(7) NodeInfo: node information including node creation time and creator information;
2.2, a two-tuple is used to represent that a total time period T of a nonlinear structure is (start, end), where start and end correspond to a start time point and an end time point of the total time period of the nonlinear structure, respectively, and there is start < end, then a total duration L of the nonlinear structure is end-start, and possible timing relationships for durations d1 and d2 of any two nodes n1 and n2 in the same nonlinear structure are as follows:
(1) neighbor (d1, d 2): n1 and n2 are in a sequential order relationship, so that d2 starts immediately after d1 ends;
(2) gradation (d1, d 2): n1 and n2 are in sequential relation but not adjacent, then d2 begins after a period of time after d1 ends;
(3) async (d1, d 2): n1 and n2 have no sequence relation, the d1 and d2 respectively execute and are not influenced by each other;
according to the time sequence relation, for the data of the five modes, firstly, displaying a three-dimensional model, then, sequentially displaying texts and pictures, simultaneously playing audio, and finally, displaying videos after the pictures and the audio are all finished so as to obtain time nodes related to the display time sequence; analyzing the sequence and the duration of the time nodes to obtain the relationship between the duration of each node of the nonlinear structure and the total time period T; finally, obtaining the logic combination flow of each modal node through conversion of time mapping;
step 3, adopting a Canvas-based layered rendering method, integrating three-dimensional scenes of a WebRTC video stream and a WebGL and webpage DOM elements, and realizing the layered rendering of a Web AR page and the mixed presentation of a nonlinear structure, wherein the Canvas-based layered implementation method comprises the following specific steps:
3.1, creating a < canvas > element and adding it to the DOM structure;
3.2, adding size and positioning attributes to the < canvas > element to support layering;
3.3, adding a transparency style element to the canvas to generate a transparent background;
3.4, repeating the three steps to generate a multilayer canvas;
and 3.5, binding the node data needing to be subjected to layered rendering display on different Canvas layers of the layers through the logic combination flow of each modal node of the nonlinear structure.
2. The seamless multi-modal content mix presentation method based on non-linear editing technology as claimed in claim 1, wherein: in the step 1, the multi-modal data file to be processed is converted into a binary form, and the first four hexadecimal codes of the header of the file are used for judging the accurate type of the file.
3. The method of claim 1, wherein the method comprises: in the step 1, for the data of the model, which is a modality, the parsing, storing and rendering of the three formats, namely obj, fbx and glTF, are provided.
CN202110562948.6A 2021-05-24 2021-05-24 Seamless multi-mode content mixing exhibition method based on nonlinear editing technology Active CN113254828B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110562948.6A CN113254828B (en) 2021-05-24 2021-05-24 Seamless multi-mode content mixing exhibition method based on nonlinear editing technology

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110562948.6A CN113254828B (en) 2021-05-24 2021-05-24 Seamless multi-mode content mixing exhibition method based on nonlinear editing technology

Publications (2)

Publication Number Publication Date
CN113254828A CN113254828A (en) 2021-08-13
CN113254828B true CN113254828B (en) 2022-09-16

Family

ID=77183910

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110562948.6A Active CN113254828B (en) 2021-05-24 2021-05-24 Seamless multi-mode content mixing exhibition method based on nonlinear editing technology

Country Status (1)

Country Link
CN (1) CN113254828B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103679800A (en) * 2013-11-21 2014-03-26 北京航空航天大学 System for generating virtual scenes of video images and method for constructing frame of system
CN109376857A (en) * 2018-09-03 2019-02-22 上海交通大学 A kind of multi-modal depth internet startup disk method of fusion structure and attribute information
US10665030B1 (en) * 2019-01-14 2020-05-26 Adobe Inc. Visualizing natural language through 3D scenes in augmented reality
CN111639408A (en) * 2020-05-27 2020-09-08 上海实迅网络科技有限公司 AR technology-based urban pipe network pipeline 3D model display method and system
CN111858954A (en) * 2020-06-29 2020-10-30 西南电子技术研究所(中国电子科技集团公司第十研究所) Task-oriented text-generated image network model

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103679800A (en) * 2013-11-21 2014-03-26 北京航空航天大学 System for generating virtual scenes of video images and method for constructing frame of system
CN109376857A (en) * 2018-09-03 2019-02-22 上海交通大学 A kind of multi-modal depth internet startup disk method of fusion structure and attribute information
US10665030B1 (en) * 2019-01-14 2020-05-26 Adobe Inc. Visualizing natural language through 3D scenes in augmented reality
CN111639408A (en) * 2020-05-27 2020-09-08 上海实迅网络科技有限公司 AR technology-based urban pipe network pipeline 3D model display method and system
CN111858954A (en) * 2020-06-29 2020-10-30 西南电子技术研究所(中国电子科技集团公司第十研究所) Task-oriented text-generated image network model

Also Published As

Publication number Publication date
CN113254828A (en) 2021-08-13

Similar Documents

Publication Publication Date Title
CN107835436B (en) A kind of real-time virtual reality fusion live broadcast system and method based on WebGL
CN106611435B (en) Animation processing method and device
CN111899322B (en) Video processing method, animation rendering SDK, equipment and computer storage medium
CN101179471B (en) Method and apparatus for implementing user personalized dynamic expression picture with characters
CN108959392B (en) Method, device and equipment for displaying rich text on 3D model
JP4833216B2 (en) Video production device and video production method
Kuroczyński et al. 3D models on triple paths-new pathways for documenting and visualizing virtual reconstructions
CN107610207A (en) A kind of method, apparatus and system of generation GIF pictures
US20180143741A1 (en) Intelligent graphical feature generation for user content
CN107767437B (en) Multilayer mixed asynchronous rendering method
CN113781660A (en) Method and device for rendering and processing virtual scene on line in live broadcast room
CN111179391A (en) Three-dimensional animation production method, system and storage medium
CN114339069B (en) Video processing method, video processing device, electronic equipment and computer storage medium
CN113254828B (en) Seamless multi-mode content mixing exhibition method based on nonlinear editing technology
CN102521864A (en) Method for simulating display screen effect in game
CN110996087B (en) Video display method and device
CN112734883A (en) Data processing method and device, electronic equipment and storage medium
WO2018049682A1 (en) Virtual 3d scene production method and related device
Giovannini et al. Creating Virtual Reality Using a Social Virtual Environment: Phygital Exhibition at the Museum Passion in Sordevolo
CN115250335A (en) Video processing method, device, equipment and storage medium
CN102724425B (en) A kind of method that teletext template is broadcasted
KR101199643B1 (en) System for virtual advertising space
CN111935492A (en) Live gift display and construction method based on video file
CN111899155B (en) Video processing method, device, computer equipment and storage medium
CN117370692A (en) Page display method and related equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant