CN111914760B - Online course video resource composition analysis method and system - Google Patents

Online course video resource composition analysis method and system Download PDF

Info

Publication number
CN111914760B
CN111914760B CN202010773342.2A CN202010773342A CN111914760B CN 111914760 B CN111914760 B CN 111914760B CN 202010773342 A CN202010773342 A CN 202010773342A CN 111914760 B CN111914760 B CN 111914760B
Authority
CN
China
Prior art keywords
video
slide
online course
analysis
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010773342.2A
Other languages
Chinese (zh)
Other versions
CN111914760A (en
Inventor
杨宗凯
徐晶
刘威
吴砥
徐建
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Central China Normal University
Original Assignee
Central China Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Central China Normal University filed Critical Central China Normal University
Priority to CN202010773342.2A priority Critical patent/CN111914760B/en
Publication of CN111914760A publication Critical patent/CN111914760A/en
Application granted granted Critical
Publication of CN111914760B publication Critical patent/CN111914760B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • G06F16/432Query formulation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/20Education
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Business, Economics & Management (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Multimedia (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Tourism & Hospitality (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Software Systems (AREA)
  • Educational Technology (AREA)
  • Molecular Biology (AREA)
  • Evolutionary Computation (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Databases & Information Systems (AREA)
  • Biophysics (AREA)
  • Educational Administration (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Economics (AREA)
  • Human Resources & Organizations (AREA)
  • Marketing (AREA)
  • Primary Health Care (AREA)
  • Strategic Management (AREA)
  • General Business, Economics & Management (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Electrically Operated Instructional Devices (AREA)

Abstract

The invention provides an analysis method and a system for online course video resource composition, which comprises the following steps: respectively determining video stream information and audio stream information from an online course video to be analyzed; extracting each video shot, and segmenting audio stream information to obtain each audio clip; identifying a video content style of a video shot; combining images and text in a video lens and voice texts corresponding to the audio segments, identifying whether the online course video contains case analysis, and if keyword words of the case analysis appear in the images, the text texts or the voice texts, considering that the online course video contains the case analysis; and taking the video content style of the video shot and the result of whether the online course video contains the case analysis as the analysis result of the online course video resource to be analyzed. The method carries out design element identification on the online course teaching video from multiple angles, and has great reference value and practical significance for learners to select online course platforms and maker evaluation resources.

Description

Online course video resource composition analysis method and system
Technical Field
The invention belongs to the technical field of teaching resource evaluation and application, and particularly relates to an analysis method and system for online course video resource composition.
Background
On-line courses are developing vigorously, especially the rise of large-scale open courses, and learning activities are increasingly being performed on the network. Compared with face-to-face teaching, learners mainly use video resources to complete learning. The video resources are taken as core learning resources of online courses and are incorporated into the standards of technical Specifications for education resource construction, evaluation Specifications for network courses and the like in China. However, in most of these specifications, the video resources are considered as a whole to be evaluated, for example, the total duration of the video resources, the total amount of the video resources, the corresponding relationship between the video resources and chapter knowledge points, and the like are not analyzed and discussed.
The production of teaching videos is undoubtedly also an important basis for teachers to develop teaching designs. An experienced teacher can flexibly organize and design the content and the mode of the video according to the requirements of teaching tasks, for example, theoretical teaching, case explanation, field interaction and the like are carried out by combining the characteristics of knowledge points, and different media materials are introduced according to the difficulty degree of the knowledge points. The method has the advantages that the composition condition of the video resources is analyzed, the teaching design intention of a teacher is facilitated to be understood, the knowledge elements of the teaching resources are decomposed, an important supporting function is provided for various intelligent teaching analyses, and the method can be applied to analysis and evaluation of teaching design and rapid retrieval of the teaching resources.
Disclosure of Invention
Aiming at the defects of the prior art, the invention aims to provide an analysis method and system for online course video resource composition, and aims to solve the problems of identification and evaluation of a teaching design method in a teaching video and detection and retrieval of teaching materials.
In order to achieve the above object, in a first aspect, an analysis method for online course video resource composition includes the following steps:
respectively determining video stream information and audio stream information from an online course video to be analyzed;
processing the video stream to extract each video shot, and segmenting the audio stream information to obtain each audio clip;
identifying a video content style of a video shot; the video content style comprises at least one of the following contents: slide lectures, lecturers, and others, such as no slide lectures and no lecturers;
identifying whether the online course video contains case analysis or not by combining the image and the text in the video lens and the voice text corresponding to the audio clip, and if the keyword text of the case analysis appears in the image, the text or the voice text, considering that the online course video contains the case analysis;
and taking the video content style of the video shot and the result of whether the online course video contains the case analysis as the analysis result of the online course video resource to be analyzed.
In an optional embodiment, the method further comprises the steps of:
if the video shot comprises the slide lecture, the target to be detected in the video shot also comprises the presentation type of the slide lecture; the presentation type includes at least one of the following ways: plain text, plain images, pictures, animations, and the like; the presentation type of the slide show presentation also belongs to the video content style.
In an alternative embodiment, the keyword words of the case analysis are: case, instance, case, say, for example, such as, example, or trial proof.
In an optional embodiment, the identifying the video content style of the video footage specifically includes the following steps:
comparing image change areas of all video lenses, detecting whether the change areas contain characters or not, combining peripheral outlines of all the image change areas, if the peripheral outlines of the combined image change areas are rectangular and contain character information, judging that a slide lecture exists in the video, and delimiting a slide lecture area range;
whether a lecturer exists in a video shot is analyzed in an image analysis mode;
combining the slide lecture note analysis result and the lecturer analysis result, the video shots are divided into pure slide lecture notes, pure lecturers, slide lecture notes mixed with lecturers or slide lecture notes without lecturers.
In an alternative embodiment, the presentation type of the slide lecture is specifically obtained by analyzing the following steps:
judging whether lines and characters exist in the range of a slide presentation area defined in a video shot, and if only lines exist, presenting the type of the lines as a pure image; if only the characters exist, the presentation type is pure characters; if the lines and the characters exist, the display type is image-text;
and judging whether the image frames in the same video shot have an inclusion relationship, if so, judging that the presentation type comprises animation.
In a second aspect, the present invention provides an online course video resource composition parsing system, including:
the video analysis unit is used for respectively determining video stream information and audio stream information from the online course video to be analyzed;
a shot extraction unit, configured to process the video stream to extract each video shot, and segment the audio stream information to obtain each audio segment;
the shot recognition unit is used for recognizing the video content style of the video shot; the video content style comprises at least one of the following contents: slide lecture, lecturer, and no slide lecture and no lecturer;
the case analysis unit is used for identifying whether the online course video comprises case analysis or not by combining the image and the text in the video lens and the voice text corresponding to the audio clip, and if the keyword and the text of the case analysis appear in the image, the text or the voice text, the online course video is considered to comprise the case analysis;
and the element analysis unit is used for taking the video content style of the video shot and the result of whether the online course video contains case analysis as the analysis result formed by the online course video resource to be analyzed.
In an optional embodiment, if the video footage includes a slide lecture, the target to be detected in the video footage identified by the footage identification unit further includes a presentation type of the slide lecture; the presentation type includes at least one of the following ways: pure characters, pure images, pictures and texts, animation and the like; the presentation type of the slide show presentation also belongs to the video content style.
In an alternative embodiment, the keyword words of the case analysis are: case, instance, case, say, for example, such as, example, or trial proof.
In an optional embodiment, the shot identification unit compares image change areas of all video shots, detects whether the change areas contain characters, merges peripheral outlines of all the image change areas, judges that a slide lecture exists in a video if the peripheral outlines of the merged image change areas are rectangular and contain character information, and delimits a slide lecture area range; whether a lecturer exists in a video shot is analyzed in an image analysis mode; and combining the slide lecture note analysis result and the lecturer analysis result to divide the video shots into a pure slide lecture note, a pure lecturer, a slide lecture note and a lecturer mixture or a slide lecturer without a lecturer.
In an alternative embodiment, the presentation type of the slide lecture is specifically obtained by analyzing the following steps:
judging whether lines and characters exist in the range of a slide presentation area defined in a video shot, and if only lines exist, presenting the type of the lines as a pure image; if only the characters exist, the presentation type is pure characters; if the lines and the characters exist, the display type is image-text;
and judging whether the image frames in the same video shot have an inclusion relationship, if so, judging that the presentation type comprises animation.
Generally, compared with the prior art, the above technical solution conceived by the present invention has the following beneficial effects:
the invention provides an analysis method and system for online course video resource composition, which can identify design elements of a teaching video from multiple dimensions such as images, voice and the like, is beneficial to understanding the teaching design intention of teachers, decomposes knowledge elements of teaching resources, and has great reference value and practical significance for analysis and evaluation of teaching design, quick retrieval of teaching resources and the like. The online course video producer can also use the invention to carry out autonomous evaluation on the produced online course video, determine the video shortage and modify the video. The online course platform can automatically complete the design element calibration of video resources through the invention, and marks on the platform for learners to select resources suitable for the learners.
Drawings
FIG. 1 is a flowchart of a method for parsing the composition of an online course video resource according to an embodiment of the present invention;
fig. 2 is an architecture diagram of a parsing system for online lesson video resource composition according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
The invention aims to meet the requirements of the teaching analysis and design field, and identifies the resource elements of the online course video and analyzes the organization time sequence of each resource element so as to obtain the composition analysis result of the video resource. The contents of the video resource elements comprise teaching slides, demonstration videos, animation materials and the like, and the organization characteristics of the resource elements comprise links such as whether a teacher goes out of the mirror or not, whether case explanation is inserted in teaching or not and the like.
The method for analyzing the course video resource composition identifies the design elements of the online course resources so as to analyze and identify the design elements from the perspective of video teaching; specifically, images in the online course video can be analyzed, whether the slide lecture is in a luxuriant picture and text, whether animation exists or not and whether a teacher goes out of the mirror are identified, whether the video contains case explanation or not is identified by combining the analysis of voice content, and then the performance of the video is evaluated.
FIG. 1 is a flow chart of a method for identifying online course video resource design elements according to an embodiment of the present invention; as shown in fig. 1, the method comprises the following steps:
s110, respectively determining video stream information and audio stream information from an online course video to be analyzed;
s120, processing the video stream to extract each video shot, and segmenting the audio stream information to obtain each audio clip;
s130, identifying the video content style of the video shot; the video content style comprises at least one of the following contents: slide lecture, lecturer, and no slide lecture and no lecturer;
s140, identifying whether the online course video contains case analysis or not by combining the image and the text in the video lens and the voice text corresponding to the audio clip, and if the keyword text of the case analysis appears in the image, the text or the voice text, considering that the online course video contains the case analysis;
s150, taking the video content style of the video shot and the result of whether the online course video contains case analysis as the analysis result of the online course video resource to be analyzed.
Specifically, if the video footage includes a slide lecture, the video content style of the video footage further includes a presentation type of the slide lecture; the presentation type includes at least one of the following ways: plain text, plain image, plain text, and animation.
In a specific embodiment, an implementation process of the online course video resource composition analysis method provided in the embodiment of the present invention includes the following steps:
(1) acquiring a target online course video, and respectively acquiring video stream information and audio stream information from a video file;
(2) processing the video stream information to obtain shot extraction, and performing audio segment segmentation on the audio stream information;
(3) judging the type of the video shot: plain slide lectures, teacher lectures, slide lectures mixed with teacher pictures, and others (no slide lectures and no teacher pictures);
(4) further analysis of the presentation type of the slide lecture in the video footage containing the slide lecture: pure characters, pure images, pictures and texts, and animation;
(5) identifying an image text and a voice text in a video lens, and judging whether case analysis exists or not by combining the image text and the voice text;
(6) and (5) summarizing the information in the steps (3) to (5) to generate a design element evaluation report of the video resource as an analysis result of the online course video resource composition.
It is understood that the design elements contain video content style and whether case analysis exists. The type of video shot, such as plain slide lecture, teacher lecture, slide lecture mixed with teacher pictures, and others, is the video content style. Further, the presentation type of the slide show presentation may also be considered as a video content style. That is, a video content style may contain multiple elements and is not limited to only one representation.
The process of processing the video stream information in the step (2) is specifically as follows:
extracting image frames according to a unit of second, finishing the extraction of video shots by calculating the correlation degree of adjacent image frames, and recording the starting time and the ending time of each video shot;
the process of processing the audio stream information in the step (2) is specifically as follows:
dividing the audio according to the pause of the speaker voice and recording the start time and the end time of the audio segment;
the identification process of the video lens type in the step (3) specifically comprises the following steps:
comparing image change areas of all video lenses, detecting whether the change areas contain characters, merging the peripheral outlines of all the image change areas (taking the maximum peripheral outline), if the peripheral outline of the merged image change area is rectangular and contains character information, judging that a slide lecture exists in the video, and drawing a slide lecture range;
meanwhile, whether people appear in the video shot or not is analyzed through the image;
combining the slide lecture with the character recognition result, and dividing the video shot into a pure slide lecture, a pure teacher, a mixed type of the slide lecture and the teacher and the like;
the specific process of the slide presentation type in the video shot in the step (4) is as follows:
judging whether lines exist in a video shot demarcated area or not, if so, judging the area to be a graph, and if characters exist at the same time, judging the area to be a graph and a text;
judging whether the image frames in the same lens have inclusion relationship, namely, if a certain image frame contains the content (characters and graphs) of a pre-preamble image frame, judging that animation exists;
the specific process of case analysis and identification in the video lens in the step (5) is as follows:
converting the audio clips into corresponding text texts by a voice recognition technology;
segmenting words in a voice text and a slide lecture, extracting keywords, searching in full-text semantic information in a regular matching mode, and judging that case analysis exists if 'case, example, if, for example, such as example and trial proof', and the like, is searched;
the invention utilizes artificial intelligence technology to identify the design elements of the online course video resources through image and voice analysis.
In a specific embodiment, the method for identifying design elements of online course video resources and the intelligent analysis system provided by the embodiment of the present invention include: the system comprises a video acquisition module, a video preprocessing module and a video design element identification module; the video acquisition module is used for reading a video file to be detected and acquiring video stream information and audio stream information; the video preprocessing module is used for processing the video stream information and the audio stream information; the video design element recognition module is used for recognizing the design elements of the video resources according to the results of the image and voice analysis and finally generating an evaluation report.
The video acquisition module can read a target online course video and load a video file to be detected in a memory; video stream information and audio stream information are obtained from the video file, respectively.
The video preprocessing module comprises an audio processing module and a video processing module, and the audio processing module is used for processing audio stream information and converting the audio stream information into a voice text through voice recognition; the video processing module is used for processing video stream information, extracting image frames according to a unit of second, completing extraction of video fragments by calculating the correlation degree of adjacent image frames, and identifying characters in the images as image texts by character identification;
the video design element identification module comprises: the system comprises a lens type identification unit, a slide show presentation type identification module, a case analysis identification module and a video design element evaluation report generation module.
The lens type identification unit firstly identifies the change areas of different lenses, merges the peripheral outlines of the change areas, and demarcates the maximum peripheral outline shape, and if the peripheral outline shape is similar to a rectangle and characters are detected, the change areas are judged as a slide lecture; meanwhile, whether a teacher exists in the image frame is judged through face recognition; merging the slide lecture notes and the human body judgment results;
the slide lecture presentation type identification module analyzes a slide lecture area defined in the image frame, detects whether characters exist or not, detects whether lines and graphs exist or not, and combines the character detection result and the graph detection result; identifying the inclusion relationship between characters and figures in the slide lecture area of adjacent image frames of the same lens, and if the inclusion relationship exists between a certain image frame and the content of a preorder image frame, judging that the image frame is an animation;
the case analysis recognition module performs 'example' on the image text and the voice text of the video, if the example is, for example, keyword retrieval such as 'trial proof', and the like, and if the matching is hit, the case analysis recognition module judges that the video contains the explanation in the form of case analysis;
and the video resource design element evaluation report generation module summarizes the identification result to generate an evaluation report.
It can be understood that the invention is based on the video image and voice analysis technology, carries on intelligent analysis and evaluation to the online course teaching video resource from the design element angle, has great reference value and practical significance to the online course platform and the producer, and also can be convenient for the learner to find the video resource suitable for the habit.
The beneficial effects of the invention can be mainly used in the following situations:
(1) the online course video producer can use the invention to independently evaluate the produced online course video design elements, determine the video shortage and modify the video shortage.
(2) The online course platform can use the invention to identify the design elements of the video uploaded by the producer and provide the learner with label information so that the learner can quickly find the resources suitable for the learner.
To further explain the design element identification method and the intelligent analysis system for online course video resources provided by the embodiment of the invention, the following is detailed with reference to the accompanying drawings and specific examples:
in a more specific embodiment, the present example provides a design element recognition system for online lesson video facing "circuit theory". The circuit theory online course video is an mp4 file. In this embodiment, a computer is required to use the system.
The specific use steps are as follows:
s1: and clicking to start preprocessing the video, and extracting the video stream and the audio stream. The subsequent steps S2-S3 are video stream preprocessing flow, and the steps S4-S5 are audio stream preprocessing flow.
S2: and carrying out similarity comparison on the extracted video frames and carrying out deletion operation, and reserving the video frames in a picture form.
S3: and identifying and extracting image text information in the picture by adopting a CTPN + CRNN algorithm. The CTPN can effectively detect the transversely distributed characters of a complex scene, and the CRNN model is a popular image-text recognition model at present and can recognize a longer text sequence.
S4: the audio stream is segmented based on an energy analysis of the audio waveform.
S5: calling a CMUSPinx voice recognition algorithm to perform voice recognition, and recording voice text information; the cmnspinx speech recognition algorithm is one of the mainstream open source speech recognition frameworks at present, originates from the university of calkymelong, and has a model in which a plurality of speeches including mandarin, english, french, spanish, and italian can be directly used.
S6: firstly, identifying the variation areas of different lenses, merging the peripheral outlines of the variation areas, defining the maximum peripheral outline shape, judging as a slide lecture if the peripheral outline shape is approximate to a rectangle and characters are detected, and executing the step S8-the step S9, otherwise skipping the step S8-the step S9.
S7: and judging whether a teacher exists in the image frame by adopting an MTCNN face detection algorithm. The MTCNN algorithm is a method for detecting human faces by a multitask cascade convolution neural network, and is one of the best human face detectors with the effect of open source codes so far.
S8: analyzing a slide presentation area defined in the image frame, detecting whether characters exist or not, and detecting whether lines and graphs exist or not.
S9: and identifying the inclusion relation of characters and graphics in the slide show lecture areas of adjacent image frames of the same lens.
S10: the image text and the voice text of the video are analyzed, and keywords like' case, example, if, for example, proof, etc. are searched in the image and voice text information in a regular matching manner, so that the case in the video is identified.
S12: summarizing the results of the steps S6-S11, judging the type of the lens according to the results of the steps S6-S7, judging the presentation type of the slide lecture according to the results of the steps S8-S9, judging whether the case analysis form explanation is contained according to the result of the step S10, and finally generating and displaying a video resource design element evaluation report.
The online course teaching video verification method can intelligently analyze and recognize design elements of the online course teaching video, the recognition result is based on the image and voice data of the video, verification can be effectively carried out on whether teachers of the online course teaching video go out of the mirror, whether pictures and texts exist, whether animation is used, whether case analysis exists and other design elements, the recognition result is accurate, and the online course teaching video verification method has great reference value and practical significance for online course platforms and makers.
FIG. 2 is a block diagram of a system for identifying online course video asset design elements provided by an embodiment of the present invention; as shown in fig. 2, includes:
a video analysis unit 210, configured to determine video stream information and audio stream information from the online course video to be parsed, respectively;
a shot extraction unit 220, configured to process the video stream to extract each video shot, and segment the audio stream information to obtain each audio segment;
a shot recognition unit 230 for recognizing a video content style of the video shot; the video content style comprises at least one of the following contents: slide lecture, lecturer, and no slide lecture and no lecturer;
the case analysis unit 240 is configured to combine the image and the text in the video shot with the voice text corresponding to the audio segment, identify whether the online lesson video includes case analysis, and if a keyword text of the case analysis appears in the image, the text, or the voice text, consider that the online lesson video includes case analysis;
and the element analysis unit 250 is configured to use the video content style of the video shot and whether the online course video includes a result of case analysis as an analysis result formed by the online course video resource to be analyzed.
It is understood that specific functions of each unit in fig. 2 can refer to detailed descriptions in the foregoing method embodiments, and are not described herein again.
It will be understood by those skilled in the art that the foregoing is only a preferred embodiment of the present invention, and is not intended to limit the invention, and that any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the scope of the present invention.

Claims (8)

1. An analysis method for online course video resource composition is characterized by comprising the following steps:
respectively determining video stream information and audio stream information from an online course video to be analyzed;
processing the video stream to extract each video shot, and segmenting the audio stream information to obtain each audio clip;
identifying a video content style of a video shot; the video content style comprises at least one of the following contents: slide lecture, lecturer, and no slide lecture and no lecturer;
the method for identifying the video content style of the video shot specifically comprises the following steps: comparing image change areas of all video lenses, detecting whether the change areas contain characters or not, combining peripheral outlines of all the image change areas, if the peripheral outlines of the combined image change areas are rectangular and contain character information, judging that a slide lecture exists in the video, and delimiting a slide lecture area range; whether a lecturer exists in a video shot is analyzed in an image analysis mode; combining the slide lecture note analysis result and the lecturer analysis result, dividing the video shots into pure slide lecture notes, pure lecturers, slide lecture notes and lecturers mixed or slide lecture notes without lecturers;
identifying whether the online course video contains case analysis or not by combining the image and the text in the video lens and the voice text corresponding to the audio clip, and if the keyword text of the case analysis appears in the image, the text or the voice text, considering that the online course video contains the case analysis;
and taking the video content style of the video shot and the result of whether the online course video contains the case analysis as the analysis result of the online course video resource to be analyzed.
2. The analysis method according to claim 1, further comprising the steps of:
if the video shot comprises the slide lecture, the target to be detected in the video shot also comprises the presentation type of the slide lecture; the presentation type includes at least one of the following ways: plain text, plain image, plain text, and animation; the presentation type of the slide show presentation also belongs to the video content style.
3. The parsing method as claimed in claim 1 or 2, wherein the keyword words of the case analysis are: case, instance, case, say, for example, such as, example, or trial proof.
4. The parsing method of claim 1, wherein the presentation type of the slide lecture is obtained by analyzing:
judging whether lines and characters exist in the range of a slide presentation area defined in a video shot, and if only lines exist, presenting the type of the lines as a pure image; if only the characters exist, the presentation type is pure characters; if the lines and the characters exist, the display type is image-text;
and judging whether the image frames in the same video shot have an inclusion relationship, if so, judging that the presentation type comprises animation.
5. An analytic system of online course video resource composition, comprising:
the video analysis unit is used for respectively determining video stream information and audio stream information from the online course video to be analyzed;
a shot extraction unit, configured to process the video stream to extract each video shot, and segment the audio stream information to obtain each audio segment;
the shot recognition unit is used for recognizing the video content style of the video shot; the video content style comprises at least one of the following contents: slide lecture, lecturer, and no slide lecture and no lecturer; the lens identification unit compares the image change areas of all the video lenses, detects whether the change areas contain characters or not, merges the peripheral outlines of all the image change areas, judges that a slide lecture exists in the video and delimits the range of the slide lecture area if the peripheral outlines of the merged image change areas are rectangular and contain character information; whether a lecturer exists in a video shot is analyzed in an image analysis mode; and combining the slide lecture note analysis result and the lecturer analysis result to divide the video shots into a pure slide lecture note, a pure lecturer, a slide lecture note and a lecturer mixture or a slide lecturer without a lecturer;
the case analysis unit is used for identifying whether the online course video comprises case analysis or not by combining the image and the text in the video lens and the voice text corresponding to the audio clip, and if the keyword and the text of the case analysis appear in the image, the text or the voice text, the online course video is considered to comprise the case analysis;
and the element analysis unit is used for taking the video content style of the video shot and the result of whether the online course video contains case analysis as the analysis result formed by the online course video resource to be analyzed.
6. The parsing system of claim 5, wherein if the video footage includes a slide lecture, the objects to be detected in the video footage identified by the footage identification unit further include a presentation type of the slide lecture; the presentation type includes at least one of the following ways: plain text, plain image, plain text, and animation; the presentation type of the slide show presentation also belongs to the video content style.
7. The parsing system as claimed in claim 5 or 6, wherein the keyword words of the case analysis are: case, instance, case, say, for example, such as, example, or trial proof.
8. The parsing system of claim 5, wherein the presentation type of the slide lecture is analyzed by:
judging whether lines and characters exist in the range of a slide presentation area defined in a video shot, and if only lines exist, presenting the type of the lines as a pure image; if only the characters exist, the presentation type is pure characters; if the lines and the characters exist, the display type is image-text;
and judging whether the image frames in the same video shot have an inclusion relationship, if so, judging that the presentation type comprises animation.
CN202010773342.2A 2020-08-04 2020-08-04 Online course video resource composition analysis method and system Active CN111914760B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010773342.2A CN111914760B (en) 2020-08-04 2020-08-04 Online course video resource composition analysis method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010773342.2A CN111914760B (en) 2020-08-04 2020-08-04 Online course video resource composition analysis method and system

Publications (2)

Publication Number Publication Date
CN111914760A CN111914760A (en) 2020-11-10
CN111914760B true CN111914760B (en) 2021-03-30

Family

ID=73287828

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010773342.2A Active CN111914760B (en) 2020-08-04 2020-08-04 Online course video resource composition analysis method and system

Country Status (1)

Country Link
CN (1) CN111914760B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114694434B (en) * 2020-12-28 2023-12-01 康立 Video teaching course intelligent generation method and system based on deep learning
CN112989117B (en) * 2021-04-14 2021-08-13 北京世纪好未来教育科技有限公司 Video classification method and device, electronic equipment and computer storage medium
US11557218B2 (en) * 2021-06-04 2023-01-17 International Business Machines Corporation Reformatting digital content for digital learning platforms using suitability scores

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7890477B2 (en) * 2003-08-20 2011-02-15 Total Training Network, Inc. Systems and methods for providing digital content
CN102612707A (en) * 2009-08-03 2012-07-25 印度孟买技术研究院 System for creating a capsule representation of an instructional video
CN102741842A (en) * 2009-12-04 2012-10-17 Tivo有限公司 Multifunction multimedia device
CN104408162A (en) * 2014-12-05 2015-03-11 国家电网公司 Multimedia system for forming text indexing and multimedia processing method
CN104464769A (en) * 2013-09-18 2015-03-25 布克查克控股有限公司 Playback system for synchronised soundtracks for electronic media content
CN106097203A (en) * 2016-06-21 2016-11-09 广州伟度计算机科技有限公司 A kind of electronics for teaching process is prepared lessons system electronic methods preparing lessons
CN107240047A (en) * 2017-05-05 2017-10-10 广州盈可视电子科技有限公司 The credit appraisal procedure and device of a kind of instructional video
CN107920280A (en) * 2017-03-23 2018-04-17 广州思涵信息科技有限公司 The accurate matched method and system of video, teaching materials PPT and voice content
CN107968959A (en) * 2017-11-15 2018-04-27 广东广凌信息科技股份有限公司 A kind of knowledge point dividing method of instructional video
CN110414352A (en) * 2019-06-26 2019-11-05 深圳市容会科技有限公司 The method and relevant device of PPT the file information are extracted from video file
CN111210673A (en) * 2018-11-21 2020-05-29 阿里巴巴集团控股有限公司 Course data processing method and device, terminal equipment and computer storage medium
CN111242962A (en) * 2020-01-15 2020-06-05 中国平安人寿保险股份有限公司 Method, device and equipment for generating remote training video and storage medium
CN111429768A (en) * 2020-03-17 2020-07-17 安徽爱学堂教育科技有限公司 Knowledge point splitting and integrating method and system based on teaching recording and broadcasting

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8280158B2 (en) * 2009-10-05 2012-10-02 Fuji Xerox Co., Ltd. Systems and methods for indexing presentation videos
CN101699426B (en) * 2009-11-06 2012-02-29 上海传知信息科技发展有限公司 Document format conversion system and method
CN105243907B (en) * 2015-11-23 2018-07-27 华中师范大学 A kind of digital education resource tutoring system based on editing machine
CN106846196A (en) * 2016-12-30 2017-06-13 首都师范大学 The course display method and device of educational robot
CN107247732A (en) * 2017-05-05 2017-10-13 广州盈可视电子科技有限公司 Exercise matching process, device and a kind of recording and broadcasting system of a kind of instructional video
US10671251B2 (en) * 2017-12-22 2020-06-02 Arbordale Publishing, LLC Interactive eReader interface generation based on synchronization of textual and audial descriptors
CN109168006A (en) * 2018-09-05 2019-01-08 高新兴科技集团股份有限公司 The video coding-decoding method that a kind of figure and image coexist

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7890477B2 (en) * 2003-08-20 2011-02-15 Total Training Network, Inc. Systems and methods for providing digital content
CN102612707A (en) * 2009-08-03 2012-07-25 印度孟买技术研究院 System for creating a capsule representation of an instructional video
CN102741842A (en) * 2009-12-04 2012-10-17 Tivo有限公司 Multifunction multimedia device
CN104464769A (en) * 2013-09-18 2015-03-25 布克查克控股有限公司 Playback system for synchronised soundtracks for electronic media content
CN104408162A (en) * 2014-12-05 2015-03-11 国家电网公司 Multimedia system for forming text indexing and multimedia processing method
CN106097203A (en) * 2016-06-21 2016-11-09 广州伟度计算机科技有限公司 A kind of electronics for teaching process is prepared lessons system electronic methods preparing lessons
CN107920280A (en) * 2017-03-23 2018-04-17 广州思涵信息科技有限公司 The accurate matched method and system of video, teaching materials PPT and voice content
CN107240047A (en) * 2017-05-05 2017-10-10 广州盈可视电子科技有限公司 The credit appraisal procedure and device of a kind of instructional video
CN107968959A (en) * 2017-11-15 2018-04-27 广东广凌信息科技股份有限公司 A kind of knowledge point dividing method of instructional video
CN111210673A (en) * 2018-11-21 2020-05-29 阿里巴巴集团控股有限公司 Course data processing method and device, terminal equipment and computer storage medium
CN110414352A (en) * 2019-06-26 2019-11-05 深圳市容会科技有限公司 The method and relevant device of PPT the file information are extracted from video file
CN111242962A (en) * 2020-01-15 2020-06-05 中国平安人寿保险股份有限公司 Method, device and equipment for generating remote training video and storage medium
CN111429768A (en) * 2020-03-17 2020-07-17 安徽爱学堂教育科技有限公司 Knowledge point splitting and integrating method and system based on teaching recording and broadcasting

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
"MOOC教学视频的优化设计研究——以美国课程中央网站Top20 MOOC为案例";王雪 等;《中国远程教育》;20180514(第5期);45-54 *
"Usability design for video lectures";Konstantinos 等;《11th European Conference on Interactive TV and Video》;20130630;163–164 *

Also Published As

Publication number Publication date
CN111914760A (en) 2020-11-10

Similar Documents

Publication Publication Date Title
CN111898441B (en) Online course video resource content identification and evaluation method and intelligent system
CN111914760B (en) Online course video resource composition analysis method and system
CN110245259B (en) Video labeling method and device based on knowledge graph and computer readable medium
Hong et al. Dynamic captioning: video accessibility enhancement for hearing impairment
US6578040B1 (en) Method and apparatus for indexing of topics using foils
Kurzhals et al. Visual movie analytics
Yang et al. In-classroom learning analytics based on student behavior, topic and teaching characteristic mining
CN112201228A (en) Multimode semantic recognition service access method based on artificial intelligence
Sun et al. Student Class Behavior Dataset: a video dataset for recognizing, detecting, and captioning students’ behaviors in classroom scenes
Chatila et al. Integrated planning and execution control of autonomous robot actions
Somandepalli et al. Computational media intelligence: Human-centered machine analysis of media
Biswas et al. Mmtoc: A multimodal method for table of content creation in educational videos
Zhao et al. A new visual interface for searching and navigating slide-based lecture videos
Hassani et al. LVTIA: A new method for keyphrase extraction from scientific video lectures
Karjus Machine-assisted mixed methods: augmenting humanities and social sciences with artificial intelligence
Navarrete et al. A review on recent advances in video-based learning research: Video features, interaction, tools, and technologies
Xu et al. Semantic navigation of powerpoint-based lecture video for autonote generation
AlMousa et al. Nlp-enriched automatic video segmentation
Li et al. Creating MAGIC: System for generating learning object metadata for instructional content
Naert et al. Per channel automatic annotation of sign language motion capture data
Vinciarelli et al. Application of information retrieval technologies to presentation slides
CN116910302A (en) Multi-mode video content effectiveness feedback visual analysis method and system
Liu et al. MND: A New Dataset and Benchmark of Movie Scenes Classified by Their Narrative Function
Kate et al. An approach for automated video indexing and video search in large lecture video archives
Manuvinakurike et al. Real-time understanding of complex discriminative scene descriptions

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant