CN115688863A - Depth knowledge tracking method based on residual connection and student near-condition feature fusion - Google Patents
Depth knowledge tracking method based on residual connection and student near-condition feature fusion Download PDFInfo
- Publication number
- CN115688863A CN115688863A CN202211354947.3A CN202211354947A CN115688863A CN 115688863 A CN115688863 A CN 115688863A CN 202211354947 A CN202211354947 A CN 202211354947A CN 115688863 A CN115688863 A CN 115688863A
- Authority
- CN
- China
- Prior art keywords
- data
- student
- knowledge
- students
- list
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 25
- 230000004927 fusion Effects 0.000 title claims abstract description 13
- 238000013528 artificial neural network Methods 0.000 claims abstract description 31
- 239000011159 matrix material Substances 0.000 claims description 26
- 125000004122 cyclic group Chemical group 0.000 claims description 23
- 238000012360 testing method Methods 0.000 claims description 13
- 230000006870 function Effects 0.000 claims description 9
- 230000008569 process Effects 0.000 claims description 8
- 238000004364 calculation method Methods 0.000 claims description 6
- 230000004913 activation Effects 0.000 claims description 4
- 238000007781 pre-processing Methods 0.000 claims description 4
- 238000012549 training Methods 0.000 claims description 4
- 238000006243 chemical reaction Methods 0.000 claims description 3
- 238000004140 cleaning Methods 0.000 claims description 3
- 238000005457 optimization Methods 0.000 claims description 3
- 238000012545 processing Methods 0.000 claims description 3
- 230000000306 recurrent effect Effects 0.000 abstract description 9
- 238000011161 development Methods 0.000 abstract description 3
- 230000015556 catabolic process Effects 0.000 abstract description 2
- 238000006731 degradation reaction Methods 0.000 abstract description 2
- 230000000694 effects Effects 0.000 description 3
- 230000003090 exacerbative effect Effects 0.000 description 3
- 230000011218 segmentation Effects 0.000 description 3
- 230000002776 aggregation Effects 0.000 description 2
- 238000004220 aggregation Methods 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000007405 data analysis Methods 0.000 description 1
- 238000007418 data mining Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Landscapes
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
本发明涉及一种基于残差连接和学生近况特征融合的深度知识追踪方法,属于知识追踪技术领域。本发明对深度知识追踪(DKT)模型存在的问题进行改进,深度知识追踪是利用学习者有时序的知识点答题相关数据和学习者对该知识点答题正确与否的相关数据,运用循环神经网络(RNN),来对学生未来答题情况进行预测。本发明将残差连接加入到DKT模型中,以解决网络信息退化的问题,以及通过学生近期的学习数据来增强学生最近的学习情况,有效的提高了模型预测的准确率,为知识追踪领域的发展提供可行的方案。
The invention relates to a deep knowledge tracking method based on fusion of residual connection and current situation features of students, and belongs to the technical field of knowledge tracking. The present invention improves the problems existing in the Deep Knowledge Tracking (DKT) model. The Deep Knowledge Tracking is to use the time-sequential knowledge point answering data of the learner and the relevant data of the learner's answer to the knowledge point correctly or not, and use the recurrent neural network (RNN), to predict the students' future answers. The invention adds the residual connection to the DKT model to solve the problem of network information degradation, and enhances the recent learning situation of the students through the recent learning data of the students, effectively improving the accuracy of the model prediction, and is a great contribution to the field of knowledge tracking. Development offers feasible solutions.
Description
技术领域technical field
本发明涉及一种基于残差连接和学生近况特征融合的深度知识追踪方法,属于知识追踪技术领域。The invention relates to a deep knowledge tracking method based on fusion of residual connection and current situation characteristics of students, and belongs to the technical field of knowledge tracking.
背景技术Background technique
知识追踪是人工智能在教育中的重要实践之一。其主要研究任务是根据学生以往的学习数据,建立一个模型对学生未来做题情况进行预测。近年来,人们提出了优于以往所提出的知识追踪方法和传统方法的基于循环神经网络(RNN)创建的深度知识追踪(DKT)。循环神经网络对于实时预测学生做题情况有着出乎意料的优势和效果,但是经过观察和实验,发现DKT还有很大的提升空间。Knowledge tracking is one of the important practices of artificial intelligence in education. Its main research task is to establish a model to predict students' future performance based on the students' past learning data. In recent years, deep knowledge tracing (DKT) based on recurrent neural network (RNN) creation has been proposed that outperforms previously proposed knowledge tracing methods and traditional methods. The cyclic neural network has unexpected advantages and effects in predicting students' performance in real time, but after observation and experimentation, it is found that DKT still has a lot of room for improvement.
随着大数据时代的到来,人工智能在教育中的应用越来越普及,学生在学习过程中产生的数据被大量存储,计算机对数据处理的能力也大大加强。教育数据挖掘和教育数据分析的发展为学习预测的发展提供了动力。深度知识追踪就是通过对学生以往的学习数据,构建以RNN为基础的模型,对学生未来做题情况进行预测。RNN对具有序列特性的数据非常有效,它能挖掘数据中的时序信息以及语义信息,利用RNN的这种能力,从学生以往的学习数据中预测未来做题情况。为教师个性化教学提供助力,同时,也提高了学生的学习效率。With the advent of the era of big data, the application of artificial intelligence in education is becoming more and more popular. The data generated by students in the learning process is stored in large quantities, and the ability of computers to process data is also greatly enhanced. Advances in educational data mining and educational data analysis have provided impetus for the development of learning prediction. In-depth knowledge tracking is to build an RNN-based model based on students' past learning data to predict students' future performance. RNN is very effective for data with sequential characteristics. It can mine the timing information and semantic information in the data. Using this ability of RNN, it can predict the future situation of the students from the past learning data. It provides assistance for teachers' personalized teaching, and at the same time, it also improves the learning efficiency of students.
发明内容Contents of the invention
本发明要解决的技术问题是提供一种基于残差连接和学生近况特征融合的深度知识追踪方法,用以解决现有的智商追踪方法中,因为忽略其他知识点预测值的合理性,从而加剧了以后预测其他知识点时的误差问题。The technical problem to be solved by the present invention is to provide a deep knowledge tracking method based on the fusion of residual connection and students' current situation features, which is used to solve the problem of ignoring the rationality of other knowledge points' predicted values in the existing IQ tracking method, thus aggravating the After understanding the error problem when predicting other knowledge points.
本发明的技术方案是:一种基于残差连接和学生近况特征融合的深度知识追踪方法,将所需数据进行one-hot编码作为循环神经网络的输入,输入到循环神经网络每一条数据生成一个预测列表,设置一个CUT_STEP值,每个学生每在循环神经网络中运算CUT_STEP个数据,就将前面的CUT_STEP个原始数据乘以权重再相加,作为输入放到全连接层,再将全连接层的输出做为补充信息加到隐含层中继续运算,直至得到输入数据的初始预测矩阵。将初始预测矩阵的每条数据与学生近期的学习数据的和进行拼接,即对学生最近的学习情况进行增强,最终经过线性层得到预测矩阵。从预测矩阵中读取所需的预测值,然后将预测值和真实值计算损失,根据损失对模型进行优化,直至模型达到相对最优。The technical solution of the present invention is: a deep knowledge tracking method based on the fusion of residual connection and current situation characteristics of students, one-hot encoding is performed on the required data as the input of the cyclic neural network, and each piece of data input to the cyclic neural network generates a For the prediction list, set a CUT_STEP value. Every time each student calculates CUT_STEP data in the cyclic neural network, multiply the previous CUT_STEP original data by weight and add them together, put them into the fully connected layer as input, and then put the fully connected layer The output of is added to the hidden layer as supplementary information to continue the operation until the initial prediction matrix of the input data is obtained. Each piece of data in the initial prediction matrix is spliced with the sum of the students' recent learning data, that is, the recent learning situation of the students is enhanced, and finally the prediction matrix is obtained through the linear layer. Read the required prediction value from the prediction matrix, then calculate the loss between the prediction value and the real value, and optimize the model according to the loss until the model reaches a relative optimum.
具体步骤为:The specific steps are:
Step1:数据预处理,将每条原始数据处理成one-hot数据形式,每条原始数据合起来就构成one-hot矩阵,具体步骤如下:Step1: Data preprocessing, process each piece of raw data into one-hot data form, each piece of raw data is combined to form a one-hot matrix, the specific steps are as follows:
Step1.1:数据读取及清理,从学生的做题记录中读取所需特征,组成新的数据形式data,所需特征包括时间序列id、学生id、知识点id、做题正误。Step1.1: Data reading and cleaning, read the required features from the student's test record, and form a new data form data. The required features include time series id, student id, knowledge point id, and correctness of the question.
Step1.2:计算知识点种类,并形成一个知识点id组成的列表K={k1,k2,...,kl}。Step1.2: Calculate the types of knowledge points, and form a list of knowledge point ids K={k 1 , k 2 ,..., k l }.
Step1.3:根据知识点列表K形成一个字典,字典形式为{知识点id:知识点在Step1.2形成的知识点列表中的位置},数学符号表示为dict={k:e},其中k∈K,K为Step1.2中形成的知识点列表,e∈{0,1,2,...,l-1}。Step1.3: Form a dictionary according to the knowledge point list K, the dictionary form is {knowledge point id: the position of the knowledge point in the knowledge point list formed in Step1.2}, and the mathematical symbol is expressed as dict={k:e}, where k∈K, K is the list of knowledge points formed in Step1.2, e∈{0, 1, 2, ..., l-1}.
Step1.4:从data中提取出每个学生的知识点列表和对应的回答正确与否的列表,形成一个序列sequences。Step1.4: Extract the list of knowledge points of each student and the corresponding list of correct answers from the data to form a sequence of sequences.
Step1.5:将sequences转化为one-hot编码形式,并将数据处理成统一形式,即每个学生有MAX_STEP条数据,每一条one-hot数据对应学生做的一道题。Step1.5: Convert the sequences into one-hot encoded form, and process the data into a unified form, that is, each student has MAX_STEP pieces of data, and each piece of one-hot data corresponds to a question that the student has done.
Step2:从Step1处理好的数据中批次读取训练数据。Step2: Read the training data in batches from the data processed in Step1.
Step3:利用循环神经网络对学生下一次做题正误进行预测,即将Step2读取的数据放入循环神经网络中运算,产生原始预测数据。其公式为:Step3: Use the cyclic neural network to predict the correctness of the student's next question, that is, put the data read in Step 2 into the cyclic neural network for calculation, and generate the original prediction data. Its formula is:
ht=tanh(Wxxt+bx+Whh(t-1)+bh) (1)h t =tanh(W x x t +b x +W h h (t-1) +b h ) (1)
式中,ht为t时刻的隐藏状态,xt为t时刻的输入,h(t-1)为t-1时刻的隐藏状态。In the formula, h t is the hidden state at time t, x t is the input at time t, and h (t-1) is the hidden state at time t-1.
Step4:循环神经网络产生的原始预测数据,通过学生近期的学习数据来增强学生最近的学习情况,进行最终预测,具体步骤为。Step4: The original prediction data generated by the cyclic neural network is used to enhance the recent learning situation of the students through the recent learning data of the students, and make the final prediction. The specific steps are as follows.
如果学生在前面的做题情况中都表现良好,但是近期的做题情况表现的很差,那么极有可能学生近期的学习情况出现了问题。并且将影响到当前的做题情况。通过学生以往的学习数据来判断学生最近的学习情况。通过实验,当拼接学生近期的学习数据的和时,效果最佳。If the student performed well in the previous problem-solving situations, but the recent problem-solving situation performed poorly, it is very likely that the student's recent learning situation has problems. And it will affect the current problem-making situation. Judge the student's recent learning situation through the student's past learning data. Through experiments, the effect is the best when splicing the sum of the students' recent learning data.
Step4.1:将Step3产生的原始预测数据与该学生近期做题数据的和进行拼接,形成心得数据ot,其公式为:Step4.1: Splicing the original prediction data generated in Step3 and the sum of the student’s recent test data to form the experience data o t , the formula is:
式中,ot为将原始预测数据与该学生近期做题数据的和进行拼接后的数据,ht为原始预测数据,xi为该学生i时刻的one-hot做题数据,n为需要拼接的学生近期做题数据的数量。In the formula, o t is the data after splicing the original forecast data and the sum of the student’s recent test data, h t is the original forecast data, x i is the one-hot test data of the student at time i, and n is the required The number of spliced students' recent test data.
Step4.2:将Step4.1拼接后的数据ot放入全连接层fc1中,调整预测的维度,fc1的表达式为:Step4.2: Put the spliced data o t of Step4.1 into the fully connected layer fc 1 , and adjust the predicted dimension. The expression of fc 1 is:
式中,和为线性层fc1的输入和输出,为权重,为偏置。In the formula, and is the input and output of the linear layer fc 1 , is the weight, for the bias.
Step4.3:将Step4.2的输出向量放入激活函数中,将输出的每个元素控制在0到1之间,得到最终预测结果,数值转换表示为:Step4.3: Put the output vector of Step4.2 into the activation function, control each element of the output between 0 and 1, and get the final prediction result. The numerical conversion is expressed as:
Step5:在Step3所述的循环神经网络的基础上加入残差连接,提高预测准确率,具体步骤为:Step5: Add residual connections on the basis of the cyclic neural network described in Step3 to improve the prediction accuracy. The specific steps are:
在Step3循环神经网络的基础上加入残差连接提高预测准确率。由于发现在循环神经网络中的预测结果会过于追求下一步知识点答题的结果,而忽略了其他知识点预测值的合理性,从而加剧了以后预测其他知识点时的误差,所以设置一个CUT_STEP(1<CUT_STEP<MAX_STEP)值用于对一个学生的one-hot矩阵的MAX_STEP进行分割,CUT_STEP即为分段数据聚合的分段值,将分段的CUT_STEP个数据相加再经过变换加入到隐含层,进入下一次循环,以上所述即为残差连接,实现残差连接具体步骤如下:On the basis of the Step3 cyclic neural network, the residual connection is added to improve the prediction accuracy. It is found that the prediction results in the cyclic neural network will be too pursuit of the results of the next knowledge point answering questions, while ignoring the rationality of the prediction values of other knowledge points, thus exacerbating the error when predicting other knowledge points in the future, so set a CUT_STEP( 1<CUT_STEP<MAX_STEP) value is used to segment the MAX_STEP of a student's one-hot matrix, and CUT_STEP is the segmentation value of segmented data aggregation, adding the segmented CUT_STEP data and adding them to the implicit Layer, enter the next cycle, the above is the residual connection, the specific steps to realize the residual connection are as follows:
Step5.1:每个学生每在循环神经网络中运算CUT_STEP个数据,就将前面的CUT_STEP个数据乘上权重W再相加,利用残差思想将前面CUT_STEP个数据根据其重要性(权重)加起来得到一个蕴含前面CUT_STEP个数据信息的向量,其中权重W={w1,W2,...,WC},C为CUT_STEP的简写。Step5.1: Every time each student calculates CUT_STEP data in the cyclic neural network, multiply the previous CUT_STEP data by the weight W and add them together, and use the residual idea to add the previous CUT_STEP data according to their importance (weight) Get a vector containing the previous CUT_STEP data information, where the weight W={w 1 , W 2 ,...,W C }, and C is the abbreviation of CUT_STEP.
Step5.2:将Step5.1产生的蕴含前面CUT_STEP个数据信息的向量作为输入放入全连接层fc2,将Step5.1的结果转化成可以直接与隐含层相加的形式,fc2的表达式为:Step5.2: Put the vector generated by Step5.1 containing the previous CUT_STEP data information into the fully connected layer fc 2 as input, and convert the result of Step5.1 into a form that can be directly added to the hidden layer, fc 2 The expression is:
式中,和为线性层fc2的输入和输出,为权重,为偏置。In the formula, and is the input and output of the linear layer fc 2 , is the weight, for the bias.
Step5.3:再将fc2的输出加到隐含层中继续运算,再在RNN中运算CUT_STEP个数据,将前面的CUT_STEp个数据相加,重复上面步骤,t时刻隐含层变化后的表达式为:Step5.3: Then add the output of fc 2 to the hidden layer to continue the operation, then calculate CUT_STEP data in RNN, add the previous CUT_STEp data, repeat the above steps, after the hidden layer changes at time t The expression is:
式中,C为CUT_STEP的简称,shd为第d个学生one-hot数据,W指为参与计算的shd赋的权重,即{sd(i-C+1),...sdi}乘以W再求和。In the formula, C is the abbreviation of CUT_STEP, s hd is the one-hot data of the dth student, W refers to the weight assigned to s hd participating in the calculation, That is {s d(i-C+1) ,...s di } multiplied by W and summed.
Step6:读取学生回答正误的真实值和模型预测学生回答正误的预测值,具体步骤为:Step6: Read the true value of the student's correct answer and the predicted value of the model's predicted student's correct or incorrect answer. The specific steps are:
由于数据集中没有给出明确的标签,所以采用的方案是:学生的第i+1条数据是该学生第i条数据生成的预测列表的标签,由于该学生第i条数据生成的预测列表是对该学生下一步做所有知识点的预测,而学生的第i+1条数据是该学生在做某个知识点时正确与否的one-hot编码,所以从这里可以看出来学生的i+1条数据是可以从该学生第i条数据生成的预测列表中找到预测值的,具体步骤如下:Since there is no clear label given in the data set, the scheme adopted is: the i+1th piece of data of the student is the label of the prediction list generated by the i-th piece of data of the student, because the prediction list generated by the i-th piece of data of the student is The next step for the student is to predict all knowledge points, and the student's i+1th piece of data is the one-hot code of whether the student is correct or not when doing a certain knowledge point, so it can be seen from here that the student's i+ 1 piece of data can be found from the prediction list generated by the i-th piece of data of the student. The specific steps are as follows:
Step6.1:从输入的one-hot矩阵中找到该学生做的除第一个以外(即零号位置)所有知识点的位置组成位置列表。Step6.1: From the input one-hot matrix, find the positions of all the knowledge points made by the student except the first one (that is, the zero position) to form a position list.
Step6.2:根据找出的位置列表从预测列表找出对应的预测值。Step6.2: Find the corresponding predicted value from the predicted list according to the found position list.
Step6.3:从输入的one-hot矩阵中找到该学生做的除第一个以外(即零号位置)所有知识点正确与否组成的真实值列表。Step6.3: From the input one-hot matrix, find the true value list consisting of all knowledge points except the first one (that is, the zero position) that the student has done.
Step7:计算预测误差并优化网络,得到可预测学生下一次做题正确与否的模型,具体步骤为:Step7: Calculate the prediction error and optimize the network to obtain a model that can predict whether the student will do the next question correctly or not. The specific steps are:
Step7.1:根据Step6中读取的真实值和预测值,利用损失函数计算损失值。Step7.1: According to the real value and predicted value read in Step6, use the loss function to calculate the loss value.
Step7.2:根据损失值,利用优化器对网络进行优化。Step7.2: According to the loss value, use the optimizer to optimize the network.
Step7.3:因为优化并不是优化一次,所欲根据需要定义Epoch的大小,决定优化的次数,即可重复几次以上步骤。Step7.3: Because optimization is not optimized once, if you want to define the size of Epoch and determine the number of optimizations according to your needs, you can repeat the above steps several times.
Step7.4:得到训练好的可预测学生下一步做题正确与否的模型。Step7.4: Obtain a trained model that can predict whether the student will do the next question correctly or not.
现有技术中,针对有时间序列的知识点答题和对应回答正误的学生学习数据,利用循环神经网络对学生下一步答题正误进行预测,由于每一次都是对学生下一步所有知识点答题结果的预测,而学生只会在下一步做一道题,所以发现预测结果会过于追求下一步知识点答题的结果,而忽略了其他知识点预测值的合理性,从而加剧了以后预测其他知识点时的误差,为了解决这个问题,本发明在循环神经网络的基础上加入了残差连接,一起完成对学生下一步答题正误的预测。同时发现,如果学生在前面的做题情况中都表现良好,但是近期的做题情况表现的很差,那么极有可能学生近期的学习情况出现了问题。并且将影响到当前的做题情况。因此本发明通过学生近期的学习数据来增强学生最近的学习情况,并取得了不错的效果。本发明运用残差连接,有效解决了预测结果过于追求下一步知识点答题的结果,而忽略了其他知识点预测值的合理性,从而加剧了以后预测其他知识点时的误差的问题。In the existing technology, for the knowledge point answers with time series and the student learning data corresponding to the correct answers, the recurrent neural network is used to predict the correctness of the students' next answer, because each time it is the result of all the knowledge points of the students in the next step. Forecasting, and students will only do one question in the next step, so it is found that the prediction result will be too pursuit of the result of answering the next knowledge point, while ignoring the rationality of the predicted value of other knowledge points, thus exacerbating the error when predicting other knowledge points in the future , in order to solve this problem, the present invention adds a residual connection on the basis of the recurrent neural network, and together completes the prediction of the student's next answer. At the same time, it was found that if the students performed well in the previous problems, but performed poorly in the recent problems, it is very likely that the students have problems in their recent learning. And it will affect the current problem-making situation. Therefore, the present invention enhances the recent learning situation of the students through the recent learning data of the students, and achieves good results. The present invention uses the residual connection to effectively solve the problem that the prediction results are too pursuit of the answer to the next knowledge point, while ignoring the rationality of the prediction values of other knowledge points, thereby exacerbating the problem of errors when predicting other knowledge points in the future.
本发明的有益效果是:本发明对深度知识追踪(DKT)模型存在的问题进行改进,深度知识追踪是利用学习者有时序的知识点答题相关数据和学习者对该知识点答题正确与否的相关数据,运用循环神经网络(RNN),来对学生未来答题情况进行预测。本发明将残差连接加入到DKT模型中,以解决网络信息退化的问题,以及通过学生近期的学习数据来增强学生最近的学习情况,有效的提高了模型预测的准确率,为知识追踪领域的发展提供可行的方案。The beneficial effects of the present invention are: the present invention improves the problems existing in the Deep Knowledge Tracking (DKT) model, and the Deep Knowledge Tracking is to use the relevant data of the learner's time-sequenced knowledge point answer questions and the learner's information on whether the knowledge point answer is correct or not. Relevant data, using recurrent neural network (RNN), to predict students' future answers. The invention adds the residual connection to the DKT model to solve the problem of network information degradation, and enhances the recent learning situation of the students through the recent learning data of the students, effectively improving the accuracy of the model prediction, and is a great contribution to the field of knowledge tracking. Development offers feasible solutions.
附图说明Description of drawings
图1是本发明的步骤流程图;Fig. 1 is a flow chart of steps of the present invention;
图2是本发明的模型示意图。Fig. 2 is a schematic diagram of a model of the present invention.
具体实施方式Detailed ways
下面结合附图和具体实施方式,对本发明作进一步说明。The present invention will be further described below in combination with the accompanying drawings and specific embodiments.
实施例1:如图2所示,本实例以公开数据集2009-2010ASSISTment数据集为例,将每个学生的数据进行处理,然后放在搭建好的模型中对模型进行训练。具体过程包括:将数据处理成可以输入模型的数据形式,从处理好的数据中批次读取训练数据,利用循环神经网络对学生下一次做题正误进行初步预测,通过学生近期的学习数据来增强学生最近的学习情况,进行最终预测,在循环神经网络的基础上加入残差连接提高预测准确率,读取真实值和预测值,计算预测误差并优化网络,得到可预测学生下一次做题正确与否的模型。Example 1: As shown in Figure 2, this example takes the public dataset 2009-2010ASSISTment dataset as an example, processes the data of each student, and then puts it in the built model to train the model. The specific process includes: processing the data into a data form that can be input into the model, reading the training data in batches from the processed data, using the recurrent neural network to make a preliminary prediction of the correctness of the students' next question, and using the students' recent learning data to Enhance students' recent learning situation, make final predictions, add residual connections on the basis of recurrent neural networks to improve prediction accuracy, read real values and predicted values, calculate prediction errors and optimize the network, and get predictable students' next questions correct or not model.
如图1所示,一种基于残差连接和学生近况特征融合的深度知识追踪方法,具体步骤为:As shown in Figure 1, a deep knowledge tracking method based on the fusion of residual connections and students' current situation features, the specific steps are:
Step1:数据预处理。Step1: Data preprocessing.
将原始数据处理成可以直接输入模型的数据,具体步骤如下:To process the raw data into data that can be directly input into the model, the specific steps are as follows:
Step1:数据预处理,将每条原始数据处理成one-hot数据形式,具体步骤如下:Step1: Data preprocessing, processing each piece of raw data into one-hot data form, the specific steps are as follows:
Step1.1:数据读取及清理:Step1.1: Data reading and cleaning:
从学生的做题记录中读取所需特征,组成新的数据形式data。所需特征包括非按时间顺序排列的编号、学生编号、知识点编号,正确与否四个特征,其中知识点为学生需掌握的知识点,每个知识点可能对应几道题来辅助学生掌握该知识点,即每个知识点可能在一个学生的数据中出现多次。Read the required features from the student's record of doing questions to form a new data form data. The required features include the number in non-chronological order, the number of students, the number of knowledge points, and the four characteristics of correctness. Among them, the knowledge points are the knowledge points that students need to master. Each knowledge point may correspond to several questions to assist students to master This knowledge point, that is, each knowledge point may appear multiple times in a student's data.
Step1.2:计算知识点种类,并形成一个知识点id组成的列表K={k1,k2,...,kl},列表长度为123,列表K的最主要作用在于为每个知识点确定位置,以便在后面的one-hot矩阵中,仅根据one-hot矩阵中的1位置即可确定是哪个知识点。Step1.2: Calculate the types of knowledge points, and form a list K={k 1 , k 2 ,...,k l } composed of knowledge point ids, the length of the list is 123. The main function of the list K is to provide each The position of the knowledge point is determined, so that in the following one-hot matrix, which knowledge point can be determined only according to the 1 position in the one-hot matrix.
Step1.3:根据Step1.2中形成的知识点列表K形成一个字典,字典形式为{知识点编号:知识点在Step1.2形成的知识点列表中的位置},数学符号表示为dict={k:e},其中k∈K,e∈{0,1,2,...122},K为Step1.2中生成的知识点列表,122为Step1.2中知识点列表长度-1,为以后形成one-hot编码做准备。Step1.3: Form a dictionary according to the knowledge point list K formed in Step1.2, the dictionary form is {knowledge point number: the position of the knowledge point in the knowledge point list formed in Step1.2}, and the mathematical symbol is expressed as dict={ k:e}, where k∈K, e∈{0, 1, 2, ... 122}, K is the list of knowledge points generated in Step1.2, 122 is the length of the list of knowledge points in Step1.2 -1, Prepare for the formation of one-hot encoding in the future.
Step1.4:从data中提取出每个学生的知识点列表和对应的回答正确与否的列表形成一个序列sequences,具体步骤如下:Step1.4: Extract the list of knowledge points of each student and the corresponding list of correct answers from the data to form a sequence sequence. The specific steps are as follows:
Step1.4.1:从data中提取出每个学生的所有数据Sr={sr1,sr2,...,srt},Sr表示所有学生的未经处理的数据,与Step1.6.4中的Sh区别,其中srt是不等长的二维数组(不等长是因为每个学生的做题数目不等),包含第t个学生的所有数据。Step1.4.1: Extract all the data of each student from data S r = {s r1 , s r2 ,..., s rt }, S r represents the unprocessed data of all students, which is the same as in Step1.6.4 The difference between Sh h , where s rt is a two-dimensional array of unequal length (the unequal length is because each student has a different number of questions to do), containing all the data of the tth student.
Step1.4.2:从每个学生的所有数据中提取出该学生的知识点列表SKi={qij|0≤i<b,j∈N*}和对应的回答正确与否的列表SAi={aij|0≤i<b,j∈N*}(用1表示正确,用0表示错误),其中SK指学生的知识点列表,i指第i个学生,即SKi指第i个学生的知识点列表,qij表示第i个学生按时间顺序做的第j道题,b为Step1.4.1中S列表的长度,即学生的个数,在该数据集中有4163个学生。j的范围取决于对应学生的做题个数(因为每个学生的做题个数不尽相同,所以每个学生的j的范围也不相同),SA指学生的回答正误列表,i指第i个学生,即SAi指第i个学生的回答正误列表。Step1.4.2: Extract the student's knowledge point list S Ki = {q ij |0≤i<b, j∈N * } and the corresponding list of correct answers S Ai = from all the data of each student {a ij |0≤i<b, j∈N * } (1 means correct, 0 means wrong), where S K refers to the student's knowledge point list, i refers to the i-th student, that is, S Ki refers to the i-th student The list of knowledge points of each student, q ij represents the jth question done by the i-th student in chronological order, b is the length of the S list in Step1.4.1, that is, the number of students, and there are 4163 students in this data set. The range of j depends on the number of questions that the corresponding students do (because the number of questions for each student is different, so the range of j for each student is also different), S A refers to the student's list of correct and incorrect answers, and i refers to The i-th student, that is, S Ai refers to the correct and incorrect answer list of the i-th student.
Step1.4.3:所有学生的知识点列表和对应的回答正误列表组成一个序列sequences。Step1.4.3: The list of knowledge points of all students and the corresponding list of correct answers form a sequence sequences.
Step1.5:将sequences转化为one-hot编码形式,每一条one-hot数据对应学生做的一道题,具体步骤如下:Step1.5: Convert the sequences into one-hot coded form, each one-hot data corresponds to a question for the students, the specific steps are as follows:
Step1.5.1:从sequences中依此读取每个学生的知识点列表(学生的知识点列表长度记为M)和回答正确与否列表。Step1.5.1: Read the list of knowledge points of each student from the sequences accordingly (the length of the list of knowledge points of the students is recorded as M) and the list of whether the answers are correct or not.
Step1.5.2:M为学生的知识点列表长度(每个学生做题数目不同,则M的值也不同),L为所有知识点个数的两倍,即L=246,根据需要设置一个MAX_STEP为50,如果一个学生的数据个数M是50的整数倍,则该学生形成一个M*246的零矩阵,否则:Step1.5.2: M is the length of the student's knowledge point list (each student has a different number of questions, so the value of M is also different), L is twice the number of all knowledge points, that is, L=246, set a MAX_STEP as needed is 50, if a student's data number M is an integer multiple of 50, then the student forms a M*246 zero matrix, otherwise:
若M%50=PIf M%50=P
则该学生形成一个Mch*246的零矩阵,Mch的表达式如下:Then the student forms a zero matrix of M ch *246, and the expression of M ch is as follows:
Mch=M+(50-P) (1)M ch =M+(50-P) (1)
MAX_STEP的作用在于将学生的数据形状进行统一化,由于每个学生的做题数目不同,所以形成的学生知识点列表长度M也不同,导致学生形成的one-hot矩阵不尽相同,而输入模型的数据形状是需要一样的,所以设定了一个MAX_STEP来对每个学生形成的one-hot矩阵进行修正,将每个学生的one-hot矩阵的长度变为MAX_STEP的整数倍,不足的补零,以便进一步将数据统一成模型需要的输入样式。The role of MAX_STEP is to unify the data shape of students. Since each student has a different number of questions, the length M of the list of students' knowledge points is also different, resulting in different one-hot matrices formed by students. The input model The shape of the data needs to be the same, so a MAX_STEP is set to correct the one-hot matrix formed by each student, and the length of each student's one-hot matrix is changed to an integer multiple of MAX_STEP, and the insufficient one-hot matrix is filled with zeros , in order to further unify the data into the input style required by the model.
Step1.5.3:使用Step1.4.2生成的学生知识点列表SKi={qij|0≤i<b,j∈N*}中的qij(即第i个学生做的第j个题),根据Step1.3形成的字典dict找到qij在Step1.2中形成知识点列表的位置s,Step1.5.3: Use the student knowledge point list S Ki generated by Step1.4.2 q ij in {q ij |0≤i<b, j∈N * } (that is, the jth question done by the i-th student), Find the position s where qij forms the knowledge point list in Step1.2 according to the dictionary dict formed in Step1.3,
如果该知识点在对应的Step1.4.2回答正确与否列表SAi中的回答正确,则第i个学生的零矩阵的第j行的第s个位置记为1,否则该学生的零矩阵的第j行的第s+l个位置变为1,l为Step1.2中形成知识点列表的长度,重复以上步骤,直到所有学生的零矩阵变成对应的one-hot矩阵。If the knowledge point is correct in the corresponding Step1.4.2 answer list S Ai , the sth position of the jth row of the i-th student’s zero matrix is recorded as 1, otherwise the student’s zero-matrix The s+l position of the jth row becomes 1, l is the length of the knowledge point list formed in Step1.2, and the above steps are repeated until the zero matrix of all students becomes the corresponding one-hot matrix.
Step1.5.4:接下来将每个学生的one-hot矩阵M*246(或者Mch*246)变成X*50*246,X根据MAX_STEP和M自动生成,即一个学生的one-hot数据根据MAX_STEP分成多块,当成多个学生的one-hot数据,重复以上步骤,所有学生根据MAX_STEP切分之后的数据组成最后的train_data,train_data里面的数据可表示为Sh={sh1,sh2,...,shi},Sh的维度为Y*50*246,其中Y为X*50*246中所有X的和,1≤i≤Y,shi={si1,si2,...sim},其中shi为第j个one-hot学生的数据,其维度为50*246,1≤m≤50。Step1.5.4: Next, change each student's one-hot matrix M*246 (or M ch *246) to X*50*246, and X is automatically generated according to MAX_STEP and M, that is, a student's one-hot data is based on MAX_STEP is divided into multiple blocks, which are regarded as one-hot data of multiple students, and the above steps are repeated. The data of all students according to MAX_STEP segmentation forms the final train_data. The data in train_data can be expressed as Sh = {s h1 , s h2 , ..., s hi }, the dimension of S h is Y*50*246, where Y is the sum of all X in X*50*246, 1≤i≤Y, s hi ={s i1 , s i2 ,. ..s im }, where s hi is the data of the jth one-hot student, and its dimension is 50*246, 1≤m≤50.
以上步骤将data数据转换成了一个Y*50*246的one-hot矩阵。The above steps convert the data data into a one-hot matrix of Y*50*246.
Step2:从Step1处理好的数据中批次读取训练数据,设置batch大小为64,决定一次加载多少个one-hot学生数据,其维度为64*50*246。Step2: Read the training data in batches from the data processed in Step1, set the batch size to 64, and decide how many one-hot student data to load at a time, and its dimension is 64*50*246.
Step3:利用循环神经网络对学生下一次做题正误进行预测,产生原始预测数据,具体步骤为:Step3: Use the cyclic neural network to predict the correctness of the students' next question, and generate the original prediction data. The specific steps are:
将Step2读取的数据放入循环神经网络中运算,由于学生的做题数据是有时序的,即学生上一步所做的题会对学生下一步做题正确率产生影响,所以选择具有记忆性和数据共享特性的RNN循环神经网络为基础构建模型,设置input_size为246,隐含层大小(hidden_size)为10,循环层的数量为1,激活函数的选择′tanh′,初始隐含层参数随机生成,ht(t时刻的隐含状态)即会作为t时刻循环神经网络的输出,也会作为t时刻的隐含状态进入下一次循环,则t时刻的隐藏状态为:Put the data read in Step 2 into the recurrent neural network for calculation. Since the data of the students' questions is time-sequential, that is, the questions the students did in the previous step will affect the correct rate of the students' next questions, so the choice has memory Build a model based on the RNN cyclic neural network with data sharing characteristics, set the input_size to 246, the hidden layer size (hidden_size) to 10, the number of cyclic layers to 1, select the activation function 'tanh', and the initial hidden layer parameters are random Generated, h t (hidden state at time t) will be used as the output of the recurrent neural network at time t, and will enter the next cycle as the hidden state at time t, then the hidden state at time t is:
ht=tanh(Wxxt+bx+Whh(t-1)+bh) (2)h t =tanh(W x x t +b x +W h h (t-1) +b h ) (2)
式中,ht为t时刻的隐藏状态,xt为t时刻的输入,h(t-1)为t-1时刻的隐藏状态。In the formula, h t is the hidden state at time t, x t is the input at time t, and h (t-1) is the hidden state at time t-1.
Step4:循环神经网络产生的原始预测数据,通过学生近期的学习数据来增强学生最近的学习情况,进行最终预测,具体步骤为:Step4: The original prediction data generated by the cyclic neural network is used to enhance the student's recent learning data through the student's recent learning data, and make the final prediction. The specific steps are:
如果学生在前面的做题情况中都表现良好,但是近期的做题情况表现的很差,那么极有可能学生近期的学习情况出现了问题。并且将影响到当前的做题情况。通过学生以往的学习数据来判断学生最近的学习情况。通过实验,当拼接学生近期的学习数据的和时,效果最佳。If the student performed well in the previous problem-solving situations, but the recent problem-solving situation performed poorly, it is very likely that the student's recent learning situation has problems. And it will affect the current problem-making situation. Judge the student's recent learning situation through the student's past learning data. Through experiments, the effect is the best when splicing the sum of the students' recent learning data.
Step4.1:将Step3产生的原始预测数据与该学生近期做题数据的和进行拼接,形成心得数据ot,其公式为:Step4.1: Splicing the original prediction data generated in Step3 and the sum of the student’s recent test data to form the experience data o t , the formula is:
式中,ot为将原始预测数据与该学生近期做题数据的和进行拼接后的数据,ht为原始预测数据,xi为该学生i时刻的one-hot做题数据,n为需要拼接的学生近期做题数据的数量。In the formula, o t is the data after splicing the original forecast data and the sum of the student’s recent test data, h t is the original forecast data, x i is the one-hot test data of the student at time i, and n is the required The number of spliced students' recent test data.
Step4.2:将Step4.1拼接后的数据ot放入全连接层fc1中,调整预测的维度,fc1的表达式为:Step4.2: Put the spliced data o t of Step4.1 into the fully connected layer fc 1 , and adjust the predicted dimension. The expression of fc 1 is:
式中,和为线性层fc1的输入和输出,为权重,为偏置。In the formula, and is the input and output of the linear layer fc 1 , is the weight, for the bias.
Step4.3:将Step4.2的输出向量放入激活函数中,将输出的每个元素控制在0到1之间,得到最终预测结果,数值转换表示为:Step4.3: Put the output vector of Step4.2 into the activation function, control each element of the output between 0 and 1, and get the final prediction result. The numerical conversion is expressed as:
Step5:在Step3循环神经网络的基础上加入残差连接,提高预测准确率,具体步骤为:Step5: Add residual connection on the basis of Step3 cyclic neural network to improve prediction accuracy. The specific steps are:
在Step3循环神经网络的基础上加入残差连接提高预测准确率。由于发现在循环神经网络中的预测结果会过于追求下一步知识点答题的结果,而忽略了其他知识点预测值的合理性,从而加剧了以后预测其他知识点时的误差,所以设置一个CUT_STEP为10,用于对一个学生的one-hot矩阵的MAX_STEP进行分割,CUT_STEP即为分段数据聚合的分段值,将分段的10个数据相加再经过变换加入到隐含层,进入下一次循环,以上所述即为残差连接,实现残差连接具体步骤如下:On the basis of the Step3 cyclic neural network, the residual connection is added to improve the prediction accuracy. It is found that the prediction results in the cyclic neural network will pursue the results of the next knowledge point answer too much, while ignoring the rationality of the prediction values of other knowledge points, which will aggravate the error when predicting other knowledge points in the future, so set a CUT_STEP as 10. It is used to segment the MAX_STEP of a student's one-hot matrix. CUT_STEP is the segmentation value of segmented data aggregation. Add the 10 segmented data and add them to the hidden layer after transformation, and enter the next time cycle, the above is the residual connection, and the specific steps to realize the residual connection are as follows:
Step5.1:每个学生每在循环神经网络中运算10个数据,就将前面的10个数据乘上权重W再相加(因为每个数据的重要性不同),利用残差思想将前面10个数据根据其重要性(权重)加起来得到一个蕴含前面10个数据信息的向量,其中权重W={w1,W2,...,WC},权重交给网络自己学习,C为CUT_STEP的简写。Step5.1: Every time each student calculates 10 data in the cyclic neural network, multiply the previous 10 data by the weight W and add them together (because the importance of each data is different), and use the residual idea to divide the previous 10 According to their importance (weight), add up a vector containing the information of the previous 10 data, where the weight W={w 1 , W 2 ,...,W C }, the weight is given to the network to learn by itself, and C is Shorthand for CUT_STEP.
Step5.2:将Step5.1产生的蕴含前面10个数据信息的向量作为输入放入全连接层fc2,fc2的输入大小为246,输出为隐含层大小10,fc2的主要作用为将Step5.1的结果转化成可以直接与隐含层相加的形式,fc2的表达式为:Step5.2: Put the vector generated by Step5.1 containing the previous 10 data information as input into the fully connected layer fc 2 , the input size of fc 2 is 246, and the output is the hidden layer size 10. The main function of fc 2 is Convert the result of Step5.1 into a form that can be directly added to the hidden layer, the expression of fc 2 is:
式中,和为线性层fc2的输入和输出,为权重,为偏置。In the formula, and is the input and output of the linear layer fc 2 , is the weight, for the bias.
Step5.3:再将fc2的输出加到隐含层中继续运算,再在RNN中运算10个数据,将前面的10个数据相加,重复上面步骤,t时刻隐含层变化后的表达式为:Step5.3: Then add the output of fc 2 to the hidden layer to continue the operation, then calculate 10 data in the RNN, add the previous 10 data, repeat the above steps, after the hidden layer changes at time t The expression is:
式中,C为CUT_STEP的简称,Shd为第d个学生one-hot数据,W指为参与计算的shd赋的权重,即{sd(i-C+1),...sdi}乘以W再求和。In the formula, C is the abbreviation of CUT_STEP, Shd is the one-hot data of the dth student, W refers to the weight assigned to the shd involved in the calculation, That is {s d(i-C+1) ,...s di } multiplied by W and summed.
Step6:读取学生回答正误的真实值和模型预测学生回答正误的预测值,具体步骤为:Step6: Read the true value of the student's correct answer and the predicted value of the model's predicted student's correct or incorrect answer. The specific steps are:
由于数据集中没有给出明确的标签,所以采用的方案是:学生的第i+1条数据是该学生第i条数据生成的预测列表的标签,由于该学生第i条数据生成的预测列表是对该学生下一步做所有知识点的预测,而学生的第i+1条数据是该学生在做某个知识点时正确与否的one-hot编码,所以从这里可以看出来学生的i+1条数据是可以从该学生第i条数据生成的预测列表中找到预测值的,具体步骤如下:Since there is no clear label given in the data set, the scheme adopted is: the i+1th piece of data of the student is the label of the prediction list generated by the i-th piece of data of the student, because the prediction list generated by the i-th piece of data of the student is The next step for the student is to predict all knowledge points, and the student's i+1th piece of data is the one-hot code of whether the student is correct or not when doing a certain knowledge point, so it can be seen from here that the student's
Step6.1:从输入的one-hot矩阵中找到该学生做的除第一个以外(即零号位置)所有知识点的位置组成位置列表,除去第一个的原因在于:由第一条数据产生对第二条的预测,第二条产生对第三条的预测,以此类推,明显,并没有产生对第一条数据的预测,即没有办法对比它的真实值和预测值,计算误差Loss,故将其去掉。Step6.1: From the input one-hot matrix, find the positions of all knowledge points made by the student except the first one (that is, the zero position) to form a position list. The reason for removing the first one is: from the first piece of data Generate a prediction for the second item, the second item generates a prediction for the third item, and so on, obviously, there is no prediction for the first item of data, that is, there is no way to compare its real value and predicted value, and calculate the error Loss, so remove it.
Step6.2:根据Step6.1中找出的位置列表从预测列表找出对应的预测值(因为每一次预测都是对该同学下一次做所有知识点的正确概率的预测,但下一次该同学做的是一个知识点,所以只需要这一个知识点的预测值即可,其他的舍弃)组成预测列表。Step6.2: According to the location list found in Step6.1, find out the corresponding prediction value from the prediction list (because each prediction is the prediction of the correct probability of the student doing all the knowledge points next time, but the next time the student What is done is a knowledge point, so only the predicted value of this knowledge point is needed, and the others are discarded) to form a prediction list.
Step6.3:从输入的one-hot矩阵中找到该学生做的除第一个以外(即零号位置)所有知识点正确与否组成的真实值列表。Step6.3: From the input one-hot matrix, find the true value list consisting of all knowledge points except the first one (that is, the zero position) that the student has done.
Step7:计算预测误差并优化网络,得到可预测学生下一次做题正确与否的模型,具体步骤为:Step7: Calculate the prediction error and optimize the network to obtain a model that can predict whether the student will do the next question correctly or not. The specific steps are:
Step7.1:根据Step6中读取的真实值和预测值,利用损失函数计算损失值。Step7.1: According to the real value and predicted value read in Step6, use the loss function to calculate the loss value.
Step7.2:根据损失值,利用优化器对网络进行优化。Step7.2: According to the loss value, use the optimizer to optimize the network.
Step7.3:根据需要定义Epoch的大小为70,决定训练几轮,即可重复几次以上步骤。Step7.3: Define the size of Epoch as 70 according to the needs, decide how many rounds to train, and repeat the above steps several times.
Step7.4:得到训练好的可预测学生下一步做题正确与否的模型。Step7.4: Obtain a trained model that can predict whether the student will do the next question correctly or not.
以上结合附图对本发明的具体实施方式作了详细说明,但是本发明并不限于上述实施方式,在本领域普通技术人员所具备的知识范围内,还可以在不脱离本发明宗旨的前提下作出各种变化。The specific embodiments of the present invention have been described in detail above in conjunction with the accompanying drawings, but the present invention is not limited to the above embodiments. Variations.
Claims (7)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211354947.3A CN115688863A (en) | 2022-11-01 | 2022-11-01 | Depth knowledge tracking method based on residual connection and student near-condition feature fusion |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211354947.3A CN115688863A (en) | 2022-11-01 | 2022-11-01 | Depth knowledge tracking method based on residual connection and student near-condition feature fusion |
Publications (1)
Publication Number | Publication Date |
---|---|
CN115688863A true CN115688863A (en) | 2023-02-03 |
Family
ID=85047127
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211354947.3A Pending CN115688863A (en) | 2022-11-01 | 2022-11-01 | Depth knowledge tracking method based on residual connection and student near-condition feature fusion |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115688863A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117474094A (en) * | 2023-12-22 | 2024-01-30 | 云南师范大学 | Knowledge tracking method based on fusion domain features of Transformer |
-
2022
- 2022-11-01 CN CN202211354947.3A patent/CN115688863A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117474094A (en) * | 2023-12-22 | 2024-01-30 | 云南师范大学 | Knowledge tracking method based on fusion domain features of Transformer |
CN117474094B (en) * | 2023-12-22 | 2024-04-09 | 云南师范大学 | Knowledge tracking method based on fusion domain features of Transformer |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110188167B (en) | An end-to-end dialogue method and system incorporating external knowledge | |
CN110428010B (en) | Knowledge tracking method | |
CN111476285B (en) | A training method for an image classification model, an image classification method, and a storage medium | |
CN112699247A (en) | Knowledge representation learning framework based on multi-class cross entropy contrast completion coding | |
CN112308326B (en) | Biological network link prediction method based on meta-path and bidirectional encoder | |
CN111538868A (en) | Knowledge Tracking Method and Exercise Recommendation Method | |
CN116402133B (en) | Knowledge graph completion method and system based on structure aggregation graph convolutional network | |
CN114398976A (en) | Machine reading comprehension method based on BERT and gated attention-enhanced network | |
Wu et al. | Self-supervised heterogeneous hypergraph network for knowledge tracing | |
CN112860847B (en) | Video question-answer interaction method and system | |
Zhou et al. | Learning with annotation of various degrees | |
CN112949929A (en) | Knowledge tracking method and system based on collaborative embedded enhanced topic representation | |
Ma et al. | [Retracted] Review on the Application of Metalearning in Artificial Intelligence | |
CN114707655A (en) | Quantum line conversion method, quantum line conversion system, storage medium and electronic equipment | |
CN116777695A (en) | Time sequence convolution knowledge tracking method for fusion project reaction | |
CN117036545A (en) | Image scene feature-based image description text generation method and system | |
CN115688863A (en) | Depth knowledge tracking method based on residual connection and student near-condition feature fusion | |
CN115544158A (en) | Multi-knowledge point dynamic knowledge tracking method applied to intelligent education system | |
Karmous‐Edwards et al. | Developing a unified definition of digital twins | |
CN116502713A (en) | Knowledge tracking method for enhancing topic similarity embedding based on weighted element path | |
Ma et al. | Dtkt: An improved deep temporal convolutional network for knowledge tracing | |
CN114911930A (en) | A Global and Local Complementary Bidirectional Attention Video Question Answering Method and System | |
CN114372151A (en) | Personalized question setting method and device, computer readable storage medium and electronic equipment | |
CN118747527A (en) | Knowledge tracing method based on knowledge concept association and historical attention information | |
CN118380050A (en) | A framework for annotation of antibiotic resistance gene attributes based on causal inference theory |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |