CN112887791A - Method for controlling video fluency - Google Patents

Method for controlling video fluency Download PDF

Info

Publication number
CN112887791A
CN112887791A CN202110091180.9A CN202110091180A CN112887791A CN 112887791 A CN112887791 A CN 112887791A CN 202110091180 A CN202110091180 A CN 202110091180A CN 112887791 A CN112887791 A CN 112887791A
Authority
CN
China
Prior art keywords
frame
video
user
teaching
generating
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110091180.9A
Other languages
Chinese (zh)
Inventor
李凯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Youlexue Technology Co ltd
Original Assignee
Shenzhen Youlexue Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Youlexue Technology Co ltd filed Critical Shenzhen Youlexue Technology Co ltd
Priority to CN202110091180.9A priority Critical patent/CN112887791A/en
Publication of CN112887791A publication Critical patent/CN112887791A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B5/00Electrically-operated educational appliances
    • G09B5/06Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
    • G09B5/065Combinations of audio and video presentations, e.g. videotapes, videodiscs, television systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments

Abstract

The invention discloses a method for controlling video fluency, which ensures that recorded and broadcast teaching videos have interactivity similar to that of live teaching and meets the requirement of on-line teaching. According to the invention, a plurality of audio, video, image, animation and other files are synthesized into the long video with a plurality of timestamps, and meanwhile, corresponding interaction rules are set for the long video, so that when a user watches recorded and broadcast teaching videos, the user can jump to corresponding video segments according to the interaction state of the user to play, transition effects among the video segments are processed, the mismatching sense of video connection is reduced, the effect of falsely playing is achieved, the user can obtain the interaction sense and experience sense of live broadcast teaching through the recorded and broadcast videos, the enthusiasm of the user for interactive question and answer is improved, the thinking of the user is exercised, and the synthesized long video can be downloaded into the terminal equipment of the user, and the influence on the teaching quality due to network environment problems is avoided.

Description

Method for controlling video fluency
Technical Field
The invention relates to the technical field of online video teaching, in particular to a method for controlling video smoothness.
Background
Along with the development of economic environment and internet technology, the appearance of online education mode has been promoted, along with the popularization of smart devices, the user acceptance of online education live broadcast scene is high, and the teaching mode of "live broadcast + instructor" can effectively improve the learning effect. Live scene of online education has stronger interactivity and on-the-spot participation sense, and interactive education will promote student's the enthusiasm of learning to improve the teaching effect. However, the participation rate of the user is low due to the limitation of time and space in the live broadcast teaching, and a lot of uncertain factors cause that the user cannot participate or cannot participate in time.
In addition to live-broadcast lectures, online education also adopts a recorded broadcast mode to carry out education. However, if online education is performed in a recorded broadcast mode, interaction between the user and the instructor cannot be realized, the classroom participation of the user is weak, supporting services are lacked after class, the teaching effect is poor, and the learning enthusiasm of the user cannot be well aroused.
Accordingly, the prior art is deficient and needs improvement.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provide a method for controlling video fluency.
The technical scheme of the invention is as follows: a method for controlling video fluency is provided, which comprises the following steps:
step 1: recording video and audio files, and making animations and pictures;
step 2: establishing a logic decision tree according to a teaching sequence, and generating a transition video for connection;
and step 3: splicing the video, the audio, the animation and the picture according to the json file of the logic decision tree to synthesize a complete long video;
and 4, step 4: acquiring interactive data of a user;
and 5: positioning to a next video starting time stamp to be played according to the interactive data of the user and the json file of the logic decision tree;
step 6: jumping to the positioned timestamp position, and playing a video clip corresponding to the interactive data of the user;
and 7: after the feedback video of the interactive data is played, reading a subsequent video clip of the video playing content before interaction, and continuing a normal teaching flow;
and 8: and when the interaction condition is met, repeating the steps 6 to 8 until the course content is completed.
Further, the specific steps of step 2 are:
step 2.1: arranging the recorded video, audio, animation and picture files according to a teaching sequence, and inserting a plurality of frames of pictures between two adjacent sections of videos to generate a transition video;
step 2.2: generating a roll call video;
step 2.3: orderly playing videos according to the teaching content, and recording a json file when the video content C1 is played and the corresponding time is T1;
step 2.4: inserting an interaction link at a time T1, setting an interaction rule, collecting interaction data of a user, and recording the interaction data to a json file;
step 2.5: analyzing and identifying the interactive data to obtain a corresponding feedback result;
step 2.6: and positioning and jumping to the corresponding video starting time stamp for playing according to the feedback result.
Further, the specific steps of step 2.1 are:
step 2.1.1: selecting the last frame A of the video A, and selecting the first frame B of the video B;
step 2.1.2: generating a frame C by the frame A and the frame B through a frame rate up-sampling frame interpolation method;
step 2.1.3: generating a frame AC by the frame A and the frame C through a frame rate up-sampling frame interpolation method;
step 2.1.4: generating a frame CB by the frame C and the frame B through a frame rate up-sampling frame interpolation method;
step 2.1.5: generating a frame A _ AC by a frame rate up-sampling frame interpolation method through the frame A and the frame AC;
step 2.1.6: generating a frame AC _ C by a frame rate up-sampling frame interpolation method for the frame AC and the frame C;
step 2.1.7: generating a frame C _ CB by the frame C and the frame CB through a frame rate up-sampling frame interpolation method;
step 2.1.8: generating a frame CB _ B by the frame rate up-sampling frame interpolation method for the frame CB and the frame B;
step 2.1.9: and splicing and merging the frame A, the frame A _ AC, the frame AC _ C, the frame C _ CB, the frame CB _ B and the frame B into the transition video according to the sequence.
Further, the specific steps of step 2.2 are:
step 2.2.1: pre-recording a video of a lecturer opening;
step 2.2.2: processing the sound position reserved for reading the name as mute;
step 2.2.3: extracting an audio file of a user name through a program;
step 2.2.4: and inserting the audio file of the user name into the sound position reserved for reading the name through ffmpeg to carry out splicing synthesis.
By adopting the scheme, the invention synthesizes a plurality of audio, video, image, animation and other files into the long video with a plurality of timestamps, and sets the corresponding interaction rules for the long video, so that when a user watches recorded and broadcast teaching video, the user can jump to the corresponding video segments according to the interaction state of the user to play, and the transition effect among the video segments is processed, the violation sense of video connection is reduced, the effect of falseness is achieved, the user can obtain the interaction sense and experience sense of live broadcast teaching through the recorded and broadcast video, the enthusiasm of the user for interactive question answering is improved, the thinking of the user is exercised, and the synthesized long video can be downloaded into the terminal equipment of the user, and the influence on the teaching quality due to network environment problems is avoided.
Drawings
FIG. 1 is a block flow diagram of the present invention.
Fig. 2 is a picture diagram of an inter-frame process.
Fig. 3 is a flow chart illustrating a frame insertion process.
Detailed Description
The invention is described in detail below with reference to the figures and the specific embodiments.
Referring to fig. 1, the present invention provides a method for controlling video fluency, comprising the following steps:
step 1: and recording video and audio files, and making animations and pictures. Aiming at different teaching contents, the pictures displayed in the videos can be videos of lecturers or animations and pictures with audio files inserted in the videos, and the corresponding videos, audios, pictures and animation files can be recorded according to the teaching process and the interactive feedback requirement so as to be edited and combined.
Step 2: and establishing a logic decision tree according to the teaching sequence, and generating a transition video for connection. Because different videos are clipped together, the problems of character jumping and blocking are easily caused at the joint of the videos, so that the videos are incoherent and unnatural. It is therefore desirable to generate transition video between two pieces of video so that the video frames are consecutive.
Step 2.1: referring to fig. 2 and fig. 3, the recorded video, audio, animation and picture files are arranged according to the teaching sequence, and a plurality of frames of pictures are inserted between two adjacent videos to generate a transition video. Specifically, the method comprises the following steps:
step 2.1.1: the last frame a of video a is selected and the first frame B of video B is selected.
Step 2.1.2: and the frame A and the frame B generate a frame C by a frame rate up-sampling frame interpolation method.
Step 2.1.3: and generating a frame AC by the frame rate up-sampling frame interpolation method for the frame A and the frame C.
Step 2.1.4: and generating a frame CB by the frame C and the frame B through a frame rate up-sampling frame interpolation method.
Step 2.1.5: and generating a frame A _ AC by the frame rate up-sampling frame interpolation method between the frame A and the frame AC.
Step 2.1.6: the frame AC and the frame C generate a frame AC _ C by a frame rate up-sampling frame interpolation method.
Step 2.1.7: and generating a frame C _ CB by the frame C and the frame CB through a frame rate up-sampling frame interpolation method.
Step 2.1.8: and the frame CB and the frame B generate a frame CB _ B by a frame rate up-sampling frame interpolation method.
Step 2.1.9: and splicing and merging the frame A, the frame A _ AC, the frame AC _ C, the frame C _ CB, the frame CB _ B and the frame B into the transition video according to the sequence.
According to the picture difference between the frame A and the frame B, a corresponding number of insertion frames are added. The insertion frame includes an insertion frame 1, an insertion frame 2, an insertion frame 3, and the like. The insertion frame 1 includes: frame C generated by frame a and frame B interpolation. The insertion frame 2 includes: frame AC generated by frame a and frame C interpolation, and frame CB generated by frame C and frame B interpolation. The insertion frame 3 includes: frame a _ AC generated from frame a and frame AC interpolation, frame AC _ C generated from frame AC and frame C interpolation, frame C _ CB generated from frame C and frame CB interpolation, and frame CB _ B generated from frame CB and frame B interpolation. By analogy, frame insertion is carried out for multiple times according to the difference between the frame A and the frame B, so that a transition video between the frame A and the frame B is generated, and the jumping feeling and the pause feeling of video connection are eliminated.
Step 2.2: and generating a roll call video. When the video is played or an interactive link is entered, the user can be called through the calling video, the interactive feeling of the user is improved, the recorded and played video can simulate the real-time interaction of live teaching, and therefore the reality is improved.
Step 2.2.1: a video of the opening opened by the instructor is prerecorded.
Step 2.2.2: the sound position where the name reading is reserved is treated as mute.
Step 2.2.3: an audio file of the user name is extracted by the program.
Step 2.2.4: and inserting the audio file of the user name into the sound position reserved for reading the name through ffmpeg to carry out splicing synthesis.
Step 2.3: and orderly playing the videos according to the teaching contents, and recording the videos to the json file when the videos are played to the video content C1 and the corresponding time is T1.
Step 2.4: and inserting an interaction link at the time T1, setting an interaction rule, collecting the interaction data of the user, and recording the interaction data to a json file.
Step 2.5: and analyzing and identifying the interactive data to obtain a corresponding feedback result.
Step 2.6: and positioning and jumping to the corresponding video starting time stamp for playing according to the feedback result.
And step 3: and splicing the video, the audio, the animation and the picture according to the json file of the logic decision tree to synthesize a complete long video.
And 4, step 4: and acquiring the interactive data of the user.
And 5: and positioning to the next video starting time stamp to be played according to the interactive data of the user and the json file of the logic decision tree.
Step 6: and jumping to the positioned timestamp position, and playing the video clip corresponding to the interactive data of the user.
And 7: and after the feedback video of the interactive data is played, reading a subsequent video clip of the video playing content before interaction, and continuing a normal teaching flow.
And 8: and when the interaction condition is met, repeating the steps 6 to 8 until the course content is completed.
In the teaching and research design stage, the lecturer content is combined, and key time points are selected from the teaching video to serve as user interactive answer points. When the video watched by the user is played to the interactive answering point, the user is prompted to answer the question or perform related operations through screen display or voice, and the audio, video and input states of the user are collected in real time in the interactive answering process of the user. And identifying and classifying according to the interaction state of the user, and skipping to the corresponding video segment for playing according to the classification feedback result, so that the user generates interactive feeling, and the learning enthusiasm of the user is aroused.
Referring to fig. 2 and fig. 3, due to the video clip, the video a and the video B are played in a linked manner, because the last frame of the video a and the first frame of the video B have great differences in the actions, expressions, and the like of the characters, the user may feel a distinct jump or pause when watching, and the watching experience of the user is affected. By collecting the frame A and the frame B, and according to the difference between the pictures, the pictures of the insertion frame 1 (frame C), the insertion frame 2 (frame AC, frame CB), the insertion frame 3 (frame A _ AC, frame AC _ C, frame C _ CB, frame CB _ B) and the like are sequentially formed, and then the pictures are spliced and combined in sequence to form a transition video, so that a smooth and natural transition is formed through the insertion frame, the pause and pause feeling generated when the videos are connected is eliminated, and the watching experience of a user is improved.
When the user watches the teaching video, the audio file in the mp3 format for collecting the user name is automatically recognized and inserted into a mute position reserved for reading the name in the video. When the video progress is played to the clip relevant to roll call, the user can be naturally and smoothly rolled, so that the interaction feeling between the user and the user is enhanced, the user has the feeling of participating in live broadcast teaching activities, and the teaching effect is improved. The roll call video can be automatically spliced in the app, so that the roll call video can be normally collected and spliced under the condition of no network connection or poor network environment.
In summary, the invention synthesizes a plurality of audio, video, image, animation and other files into a long video with a plurality of timestamps, and sets corresponding interaction rules for the long video, so that when a user watches recorded and broadcast teaching videos, the user can jump to a corresponding video segment according to the interaction state of the user to play, and the transition effect among the video segments is processed, thereby reducing the violation of video connection, achieving the effect of falsely and truly, leading the user to obtain the interaction sense and experience sense of live broadcast teaching through the recorded and broadcast videos, improving the enthusiasm of the user for interactive question and answer, training the thinking of the user, downloading the synthesized long video into the terminal equipment of the user, and avoiding the influence on the teaching quality due to network environment problems.
The present invention is not limited to the above preferred embodiments, and any modifications, equivalent substitutions and improvements made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (4)

1. A method for video fluency control, comprising the steps of:
step 1: recording video and audio files, and making animations and pictures;
step 2: establishing a logic decision tree according to a teaching sequence, and generating a transition video for connection;
and step 3: splicing the video, the audio, the animation and the picture according to the json file of the logic decision tree to synthesize a complete long video;
and 4, step 4: acquiring interactive data of a user;
and 5: positioning to a next video starting time stamp to be played according to the interactive data of the user and the json file of the logic decision tree;
step 6: jumping to the positioned timestamp position, and playing a video clip corresponding to the interactive data of the user;
and 7: after the feedback video of the interactive data is played, reading a subsequent video clip of the video playing content before interaction, and continuing a normal teaching flow;
and 8: and when the interaction condition is met, repeating the steps 6 to 8 until the course content is completed.
2. The method of claim 1, wherein the specific steps of step 2 are:
step 2.1: arranging the recorded video, audio, animation and picture files according to a teaching sequence, and inserting a plurality of frames of pictures between two adjacent sections of videos to generate a transition video;
step 2.2: generating a roll call video;
step 2.3: orderly playing videos according to the teaching content, and recording a json file when the video content C1 is played and the corresponding time is T1;
step 2.4: inserting an interaction link at a time T1, setting an interaction rule, collecting interaction data of a user, and recording the interaction data to a json file;
step 2.5: analyzing and identifying the interactive data to obtain a corresponding feedback result;
step 2.6: and positioning and jumping to the corresponding video starting time stamp for playing according to the feedback result.
3. The method for video fluency control according to claim 2, wherein the specific steps of step 2.1 are:
step 2.1.1: selecting the last frame A of the video A, and selecting the first frame B of the video B;
step 2.1.2: generating a frame C by the frame A and the frame B through a frame rate up-sampling frame interpolation method;
step 2.1.3: generating a frame AC by the frame A and the frame C through a frame rate up-sampling frame interpolation method;
step 2.1.4: generating a frame CB by the frame C and the frame B through a frame rate up-sampling frame interpolation method;
step 2.1.5: generating a frame A _ AC by a frame rate up-sampling frame interpolation method through the frame A and the frame AC;
step 2.1.6: generating a frame AC _ C by a frame rate up-sampling frame interpolation method for the frame AC and the frame C;
step 2.1.7: generating a frame C _ CB by the frame C and the frame CB through a frame rate up-sampling frame interpolation method;
step 2.1.8: generating a frame CB _ B by the frame rate up-sampling frame interpolation method for the frame CB and the frame B;
step 2.1.9: and splicing and merging the frame A, the frame A _ AC, the frame AC _ C, the frame C _ CB, the frame CB _ B and the frame B into the transition video according to the sequence.
4. The method for video fluency control according to claim 2, wherein the specific steps of step 2.2 are:
step 2.2.1: pre-recording a video of a lecturer opening;
step 2.2.2: processing the sound position reserved for reading the name as mute;
step 2.2.3: extracting an audio file of a user name through a program;
step 2.2.4: and inserting the audio file of the user name into the sound position reserved for reading the name through ffmpeg to carry out splicing synthesis.
CN202110091180.9A 2021-01-22 2021-01-22 Method for controlling video fluency Pending CN112887791A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110091180.9A CN112887791A (en) 2021-01-22 2021-01-22 Method for controlling video fluency

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110091180.9A CN112887791A (en) 2021-01-22 2021-01-22 Method for controlling video fluency

Publications (1)

Publication Number Publication Date
CN112887791A true CN112887791A (en) 2021-06-01

Family

ID=76050465

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110091180.9A Pending CN112887791A (en) 2021-01-22 2021-01-22 Method for controlling video fluency

Country Status (1)

Country Link
CN (1) CN112887791A (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103929648A (en) * 2014-03-27 2014-07-16 华为技术有限公司 Motion estimation method and device in frame rate up conversion
CN106293347A (en) * 2016-08-16 2017-01-04 广东小天才科技有限公司 The learning method of a kind of man-machine interaction and device, user terminal
CN106485964A (en) * 2016-10-19 2017-03-08 深圳市鹰硕技术有限公司 A kind of recording of classroom instruction and the method and system of program request
CN110933315A (en) * 2019-12-10 2020-03-27 Oppo广东移动通信有限公司 Image data processing method and related equipment
CN111541947A (en) * 2020-05-07 2020-08-14 天津洪恩完美未来教育科技有限公司 Teaching video processing method, device and system
CN111970562A (en) * 2020-08-17 2020-11-20 Oppo广东移动通信有限公司 Video processing method, video processing device, storage medium and electronic equipment
CN112218130A (en) * 2020-09-03 2021-01-12 北京大米科技有限公司 Control method and device for interactive video, storage medium and terminal

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103929648A (en) * 2014-03-27 2014-07-16 华为技术有限公司 Motion estimation method and device in frame rate up conversion
CN106293347A (en) * 2016-08-16 2017-01-04 广东小天才科技有限公司 The learning method of a kind of man-machine interaction and device, user terminal
CN106485964A (en) * 2016-10-19 2017-03-08 深圳市鹰硕技术有限公司 A kind of recording of classroom instruction and the method and system of program request
CN110933315A (en) * 2019-12-10 2020-03-27 Oppo广东移动通信有限公司 Image data processing method and related equipment
CN111541947A (en) * 2020-05-07 2020-08-14 天津洪恩完美未来教育科技有限公司 Teaching video processing method, device and system
CN111970562A (en) * 2020-08-17 2020-11-20 Oppo广东移动通信有限公司 Video processing method, video processing device, storage medium and electronic equipment
CN112218130A (en) * 2020-09-03 2021-01-12 北京大米科技有限公司 Control method and device for interactive video, storage medium and terminal

Similar Documents

Publication Publication Date Title
Robin Commentary: Learner-based listening and technological authenticity
US11848003B2 (en) System for communication skills training using juxtaposition of recorded takes
CN112468822B (en) Multimedia recording and broadcasting course interaction method based on video SEI message
US10741089B2 (en) Interactive immersion system for movies, television, animation, music videos, language training, entertainment, video games and social networking
CN103502969A (en) System for sequential juxtaposition of separately recorded scenes
CN112887790A (en) Method for fast interacting and playing video
CN109326154A (en) A method of human-computer interaction teaching is carried out by speech recognition engine
CN102129346A (en) Video interaction method and device
KR20190083532A (en) System for learning languages using the video selected by the learners and learning contents production method thereof
KR101066651B1 (en) Language learning method
US20130187927A1 (en) Method and System for Automated Production of Audiovisual Animations
KR101198091B1 (en) Method and system for learning contents
Roe The evolution of animated documentary
CN112887791A (en) Method for controlling video fluency
CN106993230A (en) Intelligence based on plot plays interaction control method
Duarte et al. Multimedia accessibility
JP4085015B2 (en) STREAM DATA GENERATION DEVICE, STREAM DATA GENERATION SYSTEM, STREAM DATA GENERATION METHOD, AND PROGRAM
CN108364518A (en) A kind of classroom interactions' process record method based on panorama teaching pattern
Spont Analyzing mass media through video art education: Popular pedagogy and social critique in the work of Candice Breitz
CN111080497A (en) Online programming learning client and server
CN111726692B (en) Interactive playing method of audio-video data
CN109684487A (en) Media file and its generation method and playback method
Darma et al. Motion Graphics About Successful English Presentation For Vocational Students
CN112866744A (en) Video interaction method
Kuncoro et al. Stenographic Learning Video: Design And Analysis of Video Production Using And Without Using The Script

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20210601