CN116049557A - Educational resource recommendation method based on multi-mode pre-training model - Google Patents

Educational resource recommendation method based on multi-mode pre-training model Download PDF

Info

Publication number
CN116049557A
CN116049557A CN202310097847.5A CN202310097847A CN116049557A CN 116049557 A CN116049557 A CN 116049557A CN 202310097847 A CN202310097847 A CN 202310097847A CN 116049557 A CN116049557 A CN 116049557A
Authority
CN
China
Prior art keywords
data
mode
text
training
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310097847.5A
Other languages
Chinese (zh)
Inventor
王海艳
唐瞻
骆健
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Posts and Telecommunications
Original Assignee
Nanjing University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Posts and Telecommunications filed Critical Nanjing University of Posts and Telecommunications
Priority to CN202310097847.5A priority Critical patent/CN116049557A/en
Publication of CN116049557A publication Critical patent/CN116049557A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9535Search customisation based on user profiles and personalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The invention discloses an educational resource recommending method based on a multi-mode pre-training model, which comprises the following steps: the method comprises the steps of collecting multi-mode data which are learned by a user online, carrying out corresponding preprocessing according to data of different modes, obtaining text data, picture data, audio data and numerical value data, respectively inputting the data into a single-mode receiving network of a multi-mode pre-training model, carrying out single-mode training and multi-mode cross training, carrying out mask prediction pre-training, predicting an original input shielding part by shielding a certain proportion of original input, and carrying out cross alignment training task to predict the characteristics of one mode by the specific characteristics of the other mode. The pre-trained multi-modal model can directly receive multi-modal input, input long and short memory network learning multi-modal expression vectors, project through a fully connected network, output the result as the probability that the user needs resources, and serve as recommendation ranking basis to be displayed to the user. The method and the system improve the comprehensiveness and accuracy of online education resource recommendation.

Description

Educational resource recommendation method based on multi-mode pre-training model
Technical Field
The invention relates to an educational resource recommendation method based on a multi-mode pre-training model, which can learn information of multiple modes of a user to recommend proper educational resources for the user and belongs to the technical field of educational recommendation.
Background
The rapid development of the internet has led to new data modes and industrial forms, and various data information presents geometric level growth, so that a user is difficult to obtain specific data meeting own requirements. A recommendation system aims at mining user preferences and providing possible interested items or services for users through portrait information of the users and combining historical interaction records of the users, and helping the users to retrieve related items from a large content library. The recommendation system has become a key technology for solving the overload of the online information and eliminating the information island, and is an important way for improving the quality of the information service.
With the rapid development of online curriculum technology and internet technology, large-scale open online curriculum (MOOC) has developed rapidly in recent years, attracting millions of online users. The MOOC still faces many new challenges, and high rates of coupling are one of the most serious challenges faced by MOOC. It is counted that the course completion rate of the MOOC platform is often less than 5%. For users, when a large amount of learning resources and activities are presented on the internet at the same time, the learner is inevitably confused by excessive information resources, and it is difficult to quickly find learning resources suitable for the learner. Therefore, how to reduce the learning-dropping rate and realize personalized recommendation for users is a main research problem in the field of online learning recommendation.
According to the research of the inventor, students at home and abroad develop a lot of researches aiming at the recommendation method and obtain a certain research result. Classifying from recommended objects, wherein group recommendation and personalized recommendation are performed; technically classified are recommendation methods based on collaborative filtering (Collaborative Filtering, CF), recommendation methods based on Machine Learning (ML), and recommendation methods based on Deep Learning (DL). According to the CF-based user recommendation method, resources similar to those in the user history behavior interaction records are matched and recommended to the user through similarity, or similar users are found through similarity, resources used by the similar users are recommended to the user, and the method faces the problems of data sparseness and cold start. Some research results show that the ML-based method can obtain effective recommendation results under the condition of sparse user behavior interaction data. However, the ML-based demand prediction method requires a large amount of feature processing on the data before training the model, which is time-consuming and labor-consuming. Along with the development and maturity of the deep learning technology, some students develop the research of the recommendation method based on the deep learning technology, and a certain research result is obtained. However, in the existing recommendation method based on deep learning, recommendation is generally performed based on single-mode data, and prediction accuracy needs to be improved. Unlike e-commerce/multimedia recommendations that have been developed for many years, the convergence of online educational and recommendation systems has been an initial stage and has presented a number of significant challenges. The online course platform generally only has implicit interaction behavior data (such as clicking, watching time length, collecting comments and the like of a course by students), lacks explicit scoring, and is difficult to directly play a role by a main stream recommendation algorithm.
A large number of research works show that the pre-training neural network model through massive unlabeled corpora can learn the general language representation beneficial to the downstream NLP task, and can avoid training a new model from zero. Pre-training models have been considered as an efficient strategy for training deep neural network models.
Disclosure of Invention
In order to solve the problems, the main purpose of the invention is to provide an online education based on multi-mode pre-training model education resource recommendation method, which facilitates the learning and representation of subsequent user vectors and course vectors through modeling of different interactive behavior data of multiple modes so as to improve the recommended service quality of online education resources.
In order to achieve the above purpose, the technical scheme of the invention is as follows:
in a first aspect, the present invention provides an educational resource recommendation method based on a multi-modal pre-training model, comprising:
step 1, acquiring multi-mode data in an online education system; wherein the multimodal data includes numeric class data, text class data, non-text user interaction class data, video class data, picture class data, and audio class data;
step 2, carrying out data preprocessing on the multi-mode data to obtain preprocessed text data, picture data, audio data and numerical value data;
step 3, the multi-mode pre-training model comprises a text encoder, a picture encoder, an audio encoder and a multi-mode cross encoder based on a common attention transducer;
inputting the preprocessed text data into a text encoder for masking and pre-training to obtain text vectors; inputting the preprocessed picture data into a picture encoder for masking and pre-training to obtain picture vectors; inputting the preprocessed audio data into an audio encoder for masking and pre-training to obtain an audio vector;
inputting the text vector, the picture vector and the audio vector into a multi-mode cross encoder for multi-mode alignment pre-training to obtain a multi-mode representation vector;
step 4, inputting the multi-mode expression vector and the preprocessed numerical value class data into a long-short memory network for learning to obtain a user preference vector;
step 5, inputting the user preference vector into a fully-connected network, and performing projection output to obtain the educational resource recommendation probability;
and 6, determining an educational resource recommendation result according to the educational resource recommendation probability.
In some embodiments, in step 1, the multimodal data is a series of multimodal combinations generated by a user while learning a learning resource, one multimodal combination including user operations, video, pictures, audio, text, and explicit values at a time.
In some embodiments, a method for preprocessing numeric class data includes:
the numerical class data is divided into discrete and continuous types, discrete numerical scores R,is prepared by five percent, ten percent or percentage, and is directly reserved; for the duration T of user learning u Needs to be mapped based on a threshold value, T middle For learning the median of the duration, the formula is used
Figure BDA0004072245390000031
Mapping is carried out; continuity numerical user integral S u Processing by normalization, mapping to [0,1]]Interval ∈ using formula->
Figure BDA0004072245390000041
S max The maximum value is integrated for the entire user.
In some embodiments, a method for preprocessing text class data includes: text-like data: the method comprises the steps of performing sentence segmentation on texts by using an open source tool nltk, firstly reading all files into a memory, storing the files by using list to divide the articles, then dividing the sentences by using segment, and then dividing the sentences into the files; the corpus is processed into tokens, firstly, the basic token is mainly operated as unicode conversion, punctuation segmentation, lowercase conversion, chinese character segmentation and accent symbol removal operation, and finally, the array of the words is returned; secondly, the wordpietokenizer divides each word obtained in the previous step, and the wordpiere divides words by utilizing prefix and suffix roots of the words.
In some embodiments, a method for preprocessing non-text user interaction class data includes:
according to learning time and statistics frequency, the non-text user interaction data firstly carries out independent thermal coding on various behaviors, marks the occurrence of the behaviors as 1 and marks the occurrence of the behaviors as 0, then adds up the behaviors by dividing time periods, combines the behaviors into a triplet, namely < time period, behaviors and quantity >, carries out the tokenization on the triplet, and converts the triplet into text data.
In some embodiments, the preprocessing method for video class data and picture class data comprises the following steps:
the video data are processed into pictures, and in order to extract visual characteristics according to frames, the video stream is regarded as a series of images, one video per second frame number is extracted, and the video stream is converted into picture data for processing;
the preprocessing method of the picture data comprises the following steps: scaling and regularizing the picture by utilizing a pre-trained ResNeXt-50 model, and preparing each image as three channels of RGB; three channels are converted into 2048-dimensional vectors by initializing pre-trained weights on the ImageNet dataset.
In some embodiments, a method of preprocessing audio class data includes:
the audio class data is preprocessed by using Librosa; pre-emphasis, framing, windowing preprocessing operations are performed prior to analyzing and processing the speech signals:
the pre-emphasis of the voice signal s (n) is realized by a digital filter, and the input-output relation of the pre-emphasis network is that
Figure BDA0004072245390000051
Wherein a is a pre-emphasis coefficient; taking 30ms as one frame for framing of the voice signal; the Hamming window is used for windowing, and the specific formula is as follows:
Figure BDA0004072245390000052
wherein w (N) is a voice signal subjected to Hamming window windowing, N is the order of a filter, and N represents the total length of a window function; for each short-time analysis window, obtaining a corresponding frequency spectrum through FFT; the above frequency spectrum is passed through a Mel filter bank to obtain Mel frequency spectrum; cepstrum analysis is performed on top of the Mel spectrum to obtain Mel frequency cepstrum coefficient MFCC, i.e. the characteristics of the frame of speech.
In some embodiments, step 6, determining the educational resource recommendation result according to the educational resource recommendation probability, includes:
and sequencing the educational resources according to the prediction probability from large to small, and selecting the preset number of educational resources ranked in front from the educational resources to recommend the educational resources to the user.
In a second aspect, the invention provides an educational resource recommendation device based on a multi-mode pre-training model, which comprises a processor and a storage medium;
the storage medium is used for storing instructions;
the processor is configured to operate in accordance with the instructions to perform the steps of the method according to the first aspect.
In a third aspect, the present invention provides a storage medium having stored thereon a computer program which, when executed by a processor, implements the steps of the method of the first aspect.
The beneficial effects are that: the education resource recommendation method and device based on the multi-mode pre-training model provided by the invention have the following advantages: through modeling of different interactive behavior data of multiple modes, subsequent user vectors and course vectors can be conveniently learned and represented, so that the recommended service quality of the online education resources is improved.
Drawings
FIG. 1 is a flow chart of a preferred embodiment of the educational resource recommendation method based on a multimodal pre-training model of the present invention;
FIG. 2 is a flow chart of the whole recommendation process in a preferred embodiment of the educational resource recommendation method based on the multi-modal pre-training model of the present invention;
FIG. 3 is a schematic diagram of a preferred embodiment of the educational resource recommendation method based on a multi-modal pre-training model of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more clear and clear, the present invention will be further described in detail below with reference to the accompanying drawings and examples. It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
In the description of the present invention, the descriptions of the terms "one embodiment," "some embodiments," "illustrative embodiments," "examples," "specific examples," or "some examples," etc., mean that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the present invention. In this specification, schematic representations of the above terms do not necessarily refer to the same embodiments or examples. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples.
Example 1
An educational resource recommendation method based on a multi-mode pre-training model, comprising:
step 1, acquiring multi-mode data in an online education system; wherein the multimodal data includes numeric class data, text class data, non-text user interaction class data, video class data, picture class data, and audio class data;
step 2, carrying out data preprocessing on the multi-mode data to obtain preprocessed text data, picture data, audio data and numerical value data;
step 3, the multi-mode pre-training model comprises a text encoder, a picture encoder, an audio encoder and a multi-mode cross encoder based on a common attention transducer;
inputting the preprocessed text data into a text encoder for masking and pre-training to obtain text vectors; inputting the preprocessed picture data into a picture encoder for masking and pre-training to obtain picture vectors; inputting the preprocessed audio data into an audio encoder for masking and pre-training to obtain an audio vector;
inputting the text vector, the picture vector and the audio vector into a multi-mode cross encoder for multi-mode alignment pre-training to obtain a multi-mode representation vector;
step 4, inputting the multi-mode expression vector and the preprocessed numerical value class data into a long-short memory network for learning to obtain a user preference vector;
step 5, inputting the user preference vector into a fully-connected network, and performing projection output to obtain the educational resource recommendation probability;
and 6, determining an educational resource recommendation result according to the educational resource recommendation probability.
The educational resource recommending method based on the multi-mode pre-training model comprises the following basic steps:
the method comprises the steps of obtaining original data, wherein operation, video, pictures, audio and text of a user and explicit numerical values are multi-modal combinations at a certain moment, and when the user learns a certain learning resource, a series of multi-modal combinations, namely a multi-modal combination queue, are generated.
Basic data preprocessing: and normalizing various multi-mode data and interaction records in the original input data by various methods. Wherein the multimodal data includes numeric class data, text class data, non-text interaction class data, picture class data, and audio class data.
The numerical value type data is generally an explicit score and is divided into discrete numerical values and continuous numerical values, the non-continuous numerical values are defined in a definite domain and directly reserved, and the learning duration of a specific course is required to be mapped based on a threshold value, so that the learning preference of a user to the course is reflected. The continuity value, such as the learned integral, is processed by data normalization and mapped to the [0,1] interval.
The non-numerical text is complicated text information such as course knowledge points, and more elaborate coding processing is needed, namely, all knowledge point text data are counted, the knowledge point text data are converted into vector type data which can be input by subsequent neural network processing, the word segmentation is firstly carried out, and then the vector type data are input into a text pre-training module in a multi-mode pre-training module.
The interaction data needs to be converted into numerical values, such as 'user-course' interaction behaviors of comments, collection and the like, and the occurrence of the interaction behavior can be directly recorded as 1, and the non-occurrence is recorded as 0. The time periods are then divided and accumulated to form a triplet, i.e. < time period, behavior, quantity >. And then, carrying out tokenization on the triples, and inputting the triples into an interactive type pre-training module of the multi-mode and training model.
The video class is processed into picture frames and then converted into picture class data for processing.
The picture class data is first read and converted, and the conversion includes changing brightness, contrast, saturation, enlarging target and random cutting of the picture. And resetting the size of the image, remapping the data types, and finally normalizing the image to facilitate the input of the neural network. The pre-trained ResNeXt-50 model is used for photo video. Each frame is scaled and regularized and each image is prepared as three channels of RGB. This pre-trained model can convert three channels into 2048-dimensional vectors by initializing pre-trained weights on the ImageNet dataset.
The data of the speech class are processed using Librosa, mainly extracting acoustic features including zero-crossing rate, spectral centroid, spectral attenuation, mayer frequency cepstral coefficients and chrominance frequencies for audio assessment.
Through the processing, the original input data can be converted into a data format which can be received by a subsequent neural network, training of a pre-training model is facilitated, noise information of the original data is effectively eliminated, modal differences are reduced, and effective representation of model training is obtained.
Aiming at the text interaction class, the picture video class and the voice class, respectively pre-training, encoding in different single-mode encoders and then inputting the encoded single-mode encoders into a cross transducer for multi-mode feature fusion.
The multimodal data encoded by the encoder follows the mask language modeling task in standard bert—masking about 15% of the words, image areas and audio inputs and letting the model reconstruct them given the remaining inputs. The cross pre-training then uses mainly the method of modality alignment, predicting the content of one modality by another modality.
After pre-training, the output multi-mode expression vector fuses the preference of the user to a certain part of educational resources, the user learns by using LSTM, and finally, the output classification is carried out through a fully connected network to obtain scores and display results.
In some embodiments, according to the educational resource recommendation method based on the multi-mode pre-training model in the preferred embodiment of the present invention, as shown in fig. 1 and fig. 2, the educational resource recommendation method based on the multi-mode pre-training model includes the following steps:
step 1: in an online educational system, a data collection module collects various modal data and divides the various single modal data, such as course shots, comments, barrages, course scores, notes, audio lectures, text material, and the like.
The learning record of the user in one resource is shown in table 1:
TABLE 1
2022-12-01 8:00 Open video (Limit teaching) Video addresses
2022-12-01 8:10 Pause Video frame
2022-12-01 8:22 Exit from
2022-12-01 8:25 Practice problem with opening limit Exercise question text
2022-12-01 8:40 Open the answer of exercise and exercises Answer text
2022-12-01 8:45 Uploading problem Problem picture
2022-12-01 8:47 Forum comments Comment text
…… …… ……
Other types of data formats are shown in table 2:
TABLE 2
Video scoring 4.5
Difficulty of exercises 3
User viewing duration 6:20:00
User learning points 65
And after data collection, a training data set is manufactured, data preprocessing is carried out, and data preprocessing is carried out on data of different modes.
Step 2: for explicit numerical data, the data is divided into discrete and continuous types, and the discrete numerical score R, which is generally five-component, ten-component or one-percent, can be directly reserved. For a pair ofFor a duration T of user learning u Needs to be mapped based on a threshold value, T middle Is the median of the learning duration. Using the formula
Figure BDA0004072245390000101
Mapping is performed. Continuity value such as user integral S u Treatment by normalization, using the formula +.>
Figure BDA0004072245390000102
S max The maximum value is integrated for the entire user.
Step 3: the text data such as exercises, answers, comments and barrages are characterized in that an open source tool nltk is utilized to divide texts, all files are read into a memory, list is used for dividing the texts and storing the texts, segment the sentences into the files according to the ratio of train to loss. The corpus is processed into tokens, firstly, the corpus is a basic token, the main operations are unicode conversion, punctuation mark segmentation, lowercase conversion, chinese character segmentation, accent symbol removal and other operations, and finally, the array of the words is returned; secondly, the wordbietokenizer divides each word obtained in the previous step by wordbiere, and wordbiere divides words by using more common word roots such as prefix and suffix of words, for example, the "lobida rule" is decomposed into [ "lobida", "rule" ] so that the size of a dictionary becomes acceptable.
Step 4: the user interaction data is shown in table 1, firstly, performing single-hot coding on various behaviors according to learning time and statistics frequency, then dividing a small learning time period according to total learning time length, taking the week as the small learning time length, and then accumulating to obtain sequences such as < first week, opening video, 40>, < first week, fast forwarding, 68>, < first week, pause, 20>, < first week, exit, 52> … … < eighth week, and checking example questions, 18>, wherein the sequences only need word segmentation according to triples to perform word segmentation, and the corresponding sequences are converted into token ids in a dictionary.
The video class data is processed as pictures first, and in order to extract visual features on a frame-by-frame basis, the video stream is treated as a series of images and one video per second frame number is extracted. To maintain the validity and integrity of each frame, a pre-trained ResNeXt-50 model is utilized. Each frame is scaled and regularized and each image is prepared as three channels of RGB. This pre-trained model can convert three channels into 2048-dimensional vectors by initializing pre-trained weights on the ImageNet dataset. After extraction, the video frame sequence is transformed into 2048 frame sequence-size frame features by a ResNeXt-50 model. The average of these feature vectors is used to represent visual information of the entire video. Thus, videos having different frames can be converted into the same 2048-dimensional feature vector.
Step 5: audio class data is pre-processed with Librosa for audio signals, which must be pre-emphasized, framed, windowed, etc., prior to analysis and processing of the audio signals. The purpose of these operations is to eliminate the impact on the quality of the speech signal due to aliasing, higher harmonic distortion, high frequencies, etc. caused by the human vocal organ itself and by the equipment that collects the speech signal. The method ensures that the signals obtained by the subsequent voice processing are more uniform and smoother as far as possible, provides high-quality parameters for signal parameter extraction, and improves the voice processing quality. The pre-emphasis of the voice signal s (n) is realized by a digital filter, and the input-output relation of the pre-emphasis network is that
Figure BDA0004072245390000111
Where a is the pre-emphasis coefficient, the method takes a=0.9375; taking 30ms as one frame for framing of the voice signal; the method is used for windowing by using a Hamming window immediately after the windowing, and the specific formula is as follows:
Figure BDA0004072245390000112
wherein w (N) is a voice signal subjected to Hamming window windowing, N is the order of a filter, and N represents the total length of a window function; for each short-time analysis window, obtaining a corresponding frequency spectrum through FFT; the above frequency spectrum is passed through a Mel filter bank to obtain Mel frequency spectrum; cepstrum analysis (taking the logarithm, making inverse transformation, the actual inverse transformation is generally realized by DCT discrete cosine transformation, taking the 2 nd to 13 th coefficients after DCT as MFCC coefficients) is carried out on the Mel frequency spectrum, and Mel frequency cepstrum coefficient MFCC is obtained, which is the characteristic of the frame of voice.
After preprocessing each single-mode data, inputting a multi-mode pre-training model for pre-training.
Step 6: the pre-trained input consists of three parallel BERT style models, running on the image area, text segment and audio, respectively. Each stream is made up of a series of transducer blocks (TRMs) and a new common attention transducer layer (Co-TRMs) to enable information exchange between modalities. Given an image I represented as a set of region features v 1 ,v 2 ,...v τ Text input w 1 ,w 2 ,...w t And audio input s 1 ,s 2 ,...s n The model output is ultimately represented as h v1 ,h v2 ,...h ,h w1 ,h w2 ,...h wT And h s1 ,h s2 ,...h sn . The exchange between modality streams is limited to only between specific layers and the individual modality features interact with more processing before, i.e. the selected visual features are already quite advanced and require a comparatively limited context aggregation compared to words in sentences. Given intermediate vision, language and audio representations at the joint attention TRM layer
Figure BDA0004072245390000121
And->
Figure BDA0004072245390000122
The module calculates Q, K and V matrices as in a standard transducer block. However, K and V for each modality are passed as inputs to the multi-headed attention block for the other modality.
Step 7: the first pre-training task, the mask language modeling task, is performed after the pre-training model is entered, following the mask language modeling task in standard bert—masking 15% of the word and image region inputs and letting the model reconstruct them given the remaining inputs. In the whole process, 90% of the probability of the image features of the covered image area is blocked, and 10% of the probability remains unchanged. The masking text input is handled in the same way as BERT. The model does not directly regression mask feature values, but predicts semantic class distributions for the corresponding image regions. The output of the pre-trained object detection model is used as the real label. The aim is to minimize the KL divergence between these two distributions.
Step 8: the second pre-training task is a multi-modal alignment task, with the model presented as image-text-audio pairs: and it must be predicted whether the image, text and audio are aligned, i.e.: whether the text describes an image, whether the audio illustrates the content of the image. Will output h IMG 、h CLS And h AUD As an overall representation of visual, text, and language inputs. The global representation is calculated as h using a cross-transducer structure IMG 、h CLS And h AUD The elements in between are multiplied one by one and a fully connected layer is added to predict whether the image, text and audio are aligned because the dataset only includes aligned image-text-audio pairs, so to generate negative sample pairs, the image or text is randomly replaced with another, using cross entropy as a loss function.
Step 9: after pre-training, a multi-mode pre-training model is obtained, the characteristic representation of multi-mode data can be directly obtained, the multi-mode data is input into an LSTM network for learning, then a fully-connected neural network is input, and the score of the knowledge point resource is output. And (5) inputting the output of the pre-training model into a neural network for classification, and obtaining scores for recommendation.
Example 2
In a second aspect, the present embodiment provides an educational resource recommendation device based on a multi-mode pre-training model, including a processor and a storage medium;
the storage medium is used for storing instructions;
the processor is operative according to the instructions to perform the steps of the method according to embodiment 1.
Example 3
In a third aspect, the present embodiment provides a storage medium having stored thereon a computer program which, when executed by a processor, implements the steps of the method described in embodiment 1.
It will be appreciated by those skilled in the art that embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
The foregoing is only a preferred embodiment of the invention, it being noted that: it will be apparent to those skilled in the art that various modifications and adaptations can be made without departing from the principles of the present invention, and such modifications and adaptations are intended to be comprehended within the scope of the invention.

Claims (10)

1. An educational resource recommendation method based on a multi-mode pre-training model is characterized by comprising the following steps:
step 1, acquiring multi-mode data in an online education system; wherein the multimodal data includes numeric class data, text class data, non-text user interaction class data, video class data, picture class data, and audio class data;
step 2, carrying out data preprocessing on the multi-mode data to obtain preprocessed text data, picture data, audio data and numerical value data;
step 3, the multi-mode pre-training model comprises a text encoder, a picture encoder, an audio encoder and a multi-mode cross encoder based on a common attention transducer;
inputting the preprocessed text data into a text encoder for masking and pre-training to obtain text vectors; inputting the preprocessed picture data into a picture encoder for masking and pre-training to obtain picture vectors; inputting the preprocessed audio data into an audio encoder for masking and pre-training to obtain an audio vector;
inputting the text vector, the picture vector and the audio vector into a multi-mode cross encoder for multi-mode alignment pre-training to obtain a multi-mode representation vector;
step 4, inputting the multi-mode expression vector and the preprocessed numerical value class data into a long-short memory network for learning to obtain a user preference vector;
step 5, inputting the user preference vector into a fully-connected network, and performing projection output to obtain the educational resource recommendation probability;
and 6, determining an educational resource recommendation result according to the educational resource recommendation probability.
2. The method of claim 1, wherein in step 1, the multimodal data is a series of multimodal combinations generated by a user while learning a learning resource, one multimodal combination including user operations, video, pictures, audio, text, and explicit numerical values at a time.
3. The method of claim 1, wherein the preprocessing method of the numeric class data comprises:
the numerical value class data are divided into discrete type and continuous type, wherein the discrete type numerical value score R is five-point system, ten-point system or percentage system and is directly reserved; for the duration T of user learning u Needs to be mapped based on a threshold value, T middle For learning the median of the duration, the formula is used
Figure FDA0004072245380000021
Mapping is carried out; continuity numerical user integral S u Processing by normalization, mapping to [0,1]]Interval ∈ using formula->
Figure FDA0004072245380000022
S max The maximum value is integrated for the entire user.
4. The method of claim 1, wherein the preprocessing method of the text class data comprises: text-like data: the method comprises the steps of performing sentence segmentation on texts by using an open source tool nltk, firstly reading all files into a memory, storing the files by using list to divide the articles, then dividing the sentences by using segment, and then dividing the sentences into the files; the corpus is processed into tokens, firstly, the basic token is mainly operated as unicode conversion, punctuation segmentation, lowercase conversion, chinese character segmentation and accent symbol removal operation, and finally, the array of the words is returned; secondly, the wordpietokenizer divides each word obtained in the previous step, and the wordpiere divides words by utilizing prefix and suffix roots of the words.
5. The method of claim 1, wherein the preprocessing method of the non-text user interaction class data comprises:
according to learning time and statistics frequency, the non-text user interaction data firstly carries out independent thermal coding on various behaviors, marks the occurrence of the behaviors as 1 and marks the occurrence of the behaviors as 0, then adds up the behaviors by dividing time periods, combines the behaviors into a triplet, namely < time period, behaviors and quantity >, carries out the tokenization on the triplet, and converts the triplet into text data.
6. The method according to claim 1, wherein the preprocessing method of the video class data and the picture class data comprises:
the video data are processed into pictures, and in order to extract visual characteristics according to frames, the video stream is regarded as a series of images, one video per second frame number is extracted, and the video stream is converted into picture data for processing;
the preprocessing method of the picture data comprises the following steps: scaling and regularizing the picture by utilizing a pre-trained ResNeXt-50 model, and preparing each image as three channels of RGB; three channels are converted into 2048-dimensional vectors by initializing pre-trained weights on the ImageNet dataset.
7. The method of claim 1, wherein the preprocessing method of the audio class data comprises: the audio class data is preprocessed by using Librosa;
pre-emphasis, framing, windowing preprocessing operations are performed prior to analyzing and processing the speech signals:
the pre-emphasis of the voice signal s (n) is realized by a digital filter, and the input-output relation of the pre-emphasis network is that
Figure FDA0004072245380000031
Wherein a is a pre-emphasis coefficient; taking 30ms as one frame for framing of the voice signal; the Hamming window is used for windowing, and the specific formula is as follows:
Figure FDA0004072245380000032
wherein w (N) is a voice signal subjected to Hamming window windowing, N is the order of a filter, and N represents the total length of a window function; for each short-time analysis window, obtaining a corresponding frequency spectrum through FFT; the above frequency spectrum is passed through a Mel filter bank to obtain Mel frequency spectrum; cepstrum analysis is performed on top of the Mel spectrum to obtain Mel frequency cepstrum coefficient MFCC, i.e. the characteristics of the frame of speech.
8. The method of claim 1, wherein step 6 of determining the educational resource recommendation result based on the educational resource recommendation probability comprises:
and sequencing the educational resources according to the prediction probability from large to small, and selecting the preset number of educational resources ranked in front from the educational resources to recommend the educational resources to the user.
9. An educational resource recommendation device based on a multi-mode pre-training model is characterized by comprising a processor and a storage medium;
the storage medium is used for storing instructions;
the processor being operative according to the instructions to perform the steps of the method according to any one of claims 1 to 8.
10. A storage medium having stored thereon a computer program, which when executed by a processor performs the steps of the method according to any of claims 1 to 8.
CN202310097847.5A 2023-02-10 2023-02-10 Educational resource recommendation method based on multi-mode pre-training model Pending CN116049557A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310097847.5A CN116049557A (en) 2023-02-10 2023-02-10 Educational resource recommendation method based on multi-mode pre-training model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310097847.5A CN116049557A (en) 2023-02-10 2023-02-10 Educational resource recommendation method based on multi-mode pre-training model

Publications (1)

Publication Number Publication Date
CN116049557A true CN116049557A (en) 2023-05-02

Family

ID=86125456

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310097847.5A Pending CN116049557A (en) 2023-02-10 2023-02-10 Educational resource recommendation method based on multi-mode pre-training model

Country Status (1)

Country Link
CN (1) CN116049557A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116702094A (en) * 2023-08-01 2023-09-05 国家计算机网络与信息安全管理中心 Group application preference feature representation method
CN117350409A (en) * 2023-12-04 2024-01-05 环球数科集团有限公司 Man-machine dialogue model training system based on machine learning

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116702094A (en) * 2023-08-01 2023-09-05 国家计算机网络与信息安全管理中心 Group application preference feature representation method
CN116702094B (en) * 2023-08-01 2023-12-22 国家计算机网络与信息安全管理中心 Group application preference feature representation method
CN117350409A (en) * 2023-12-04 2024-01-05 环球数科集团有限公司 Man-machine dialogue model training system based on machine learning
CN117350409B (en) * 2023-12-04 2024-03-01 环球数科集团有限公司 Man-machine dialogue model training system based on machine learning

Similar Documents

Publication Publication Date Title
Zhu et al. Uncovering the temporal context for video question answering
CN110020437B (en) Emotion analysis and visualization method combining video and barrage
Lippi et al. Argument mining from speech: Detecting claims in political debates
WO2022095380A1 (en) Ai-based virtual interaction model generation method and apparatus, computer device and storage medium
CN113435203B (en) Multi-modal named entity recognition method and device and electronic equipment
Kafle et al. Evaluating the usability of automatically generated captions for people who are deaf or hard of hearing
EP3408766A1 (en) Digital media content extraction natural language processing system
CN116049557A (en) Educational resource recommendation method based on multi-mode pre-training model
Ameisen Building Machine Learning Powered Applications: Going from Idea to Product
CN113297370B (en) End-to-end multi-modal question-answering method and system based on multi-interaction attention
US11947582B2 (en) Enhanced knowledge delivery and attainment using a question answering system
CN110825867B (en) Similar text recommendation method and device, electronic equipment and storage medium
Dinkov et al. Predicting the leading political ideology of YouTube channels using acoustic, textual, and metadata information
US10083219B2 (en) Enhanced knowledge delivery and attainment using a question answering system
CN113380271B (en) Emotion recognition method, system, device and medium
CN114281948A (en) Summary determination method and related equipment thereof
CN117252259A (en) Deep learning-based natural language understanding method and AI teaching aid system
CN116980665A (en) Video processing method, device, computer equipment, medium and product
Tsujimura et al. Automatic Explanation Spot Estimation Method Targeted at Text and Figures in Lecture Slides.
CN112507115B (en) Method and device for classifying emotion words in barrage text and storage medium
CN111340329A (en) Actor assessment method and device and electronic equipment
CN114547435A (en) Content quality identification method, device, equipment and readable storage medium
CN112347786A (en) Artificial intelligence scoring training method and device
CN111081095A (en) Video and audio teaching platform, analysis subsystem and method, recommendation subsystem and method
Gala et al. Real-time cognitive evaluation of online learners through automatically generated questions

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination