Open Access
ARTICLE
Noise-Filtering Enhanced Deep Cognitive Diagnosis Model for Latent Skill Discovering
1 National Engineering Research Center of Educational Big Data, Central China Normal University, Wuhan, 430079, China
2 Faculty of Artificial Intelligence in Education, Central China Normal University, Wuhan, 430079, China
3 School of Computer Science and Artificial Intelligence, Wuhan Textile University, Wuhan, 430079, China
* Corresponding Author: Jing Geng. Email:
(This article belongs to the Special Issue: Continual Learning Techniques for Mobile-based Smart Healthcare Information Systems)
Intelligent Automation & Soft Computing 2023, 37(2), 1311-1324. https://doi.org/10.32604/iasc.2023.038481
Received 14 December 2022; Accepted 20 February 2023; Issue published 21 June 2023
A correction of this article was approved in:
Correction: Noise-Filtering Enhanced Deep Cognitive Diagnosis Model for Latent Skill Discovering
Read correction
Abstract
Educational data mining based on student cognitive diagnosis analysis can provide an important decision basis for personalized learning tutoring of students, which has attracted extensive attention from scholars at home and abroad and has made a series of important research progress. To this end, we propose a noise-filtering enhanced deep cognitive diagnosis method to improve the fitting ability of traditional models and obtain students’ skill mastery status by mining the interaction between students and problems nonlinearly through neural networks. First, modeling complex interactions between students and problems with multidimensional features based on cognitive processing theory can enhance the interpretability of the proposed model; second, the neural network is used to predict students’ learning performance, diagnose students’ skill mastery and provide immediate feedback; finally, by comparing the proposed model with several baseline models, extensive experimental results on real data sets demonstrate that the proposed Finally, by comparing the proposed model with several baseline models, the extensive experimental results on the actual data set demonstrate that the proposed model not only improves the accuracy of predicting students’ learning performance but also enhances the interpretability of the neurocognitive diagnostic model.Keywords
Personalized education for students’ characteristics in the context of big data in education in order to realize teaching according to their abilities is an important direction of current research on smart education. The rapid development of online education and major online learning platforms in recent years has promoted the large-scale popularity of online learning [1–3]. With the increasing abundance of learning resources in online learning platforms, it has become increasingly important to provide personalized learning services for students. Based on students’ online learning data, it is an important research hotspot to analyze students’ learning status by means of educational data mining (EDM) technology and make accurate recommendations for personalized learning based on students’ actual characteristics in order to realize teaching based on their abilities [4–8]. In order to achieve personalized learning guidance for students, it is first necessary to identify the current learning status of students [9–12]. Traditional teaching methods rely on teachers’ experience to make judgments, which not only takes a lot of time and energy, but also makes it difficult to guarantee the accuracy of evaluation results, and is not suitable for application scenarios where the size of students increases dramatically in the online learning environment [13–15].
To this end, the researcher attempted to introduce the cognitive diagnosis (CD) method in educational psychology to characterize students’ learning status [16–18]. The key to cognitive diagnosis is to construct a Q matrix reflecting students’ problem solving process and to choose appropriate cognitive diagnostic models (CD models) to model students [19,20]. The cognitive diagnostic model (CDM) [21,22] is an important tool for achieving cognitive diagnosis, and an increasing number of researchers have worked on the development of this model. The traditional CDM uses different item response functions to probabilistically model students’ response processes through different learning assumptions to diagnose their skill mastery status [23,24]. However, traditional CDM has some problems, such as the convergence effect of the model is highly dependent on the sample size, number of skills and number of parameters, which cannot handle sparse data, and it is difficult to explore the potential nonlinear interactions between students and test questions.
With the gradual development of computer technology, deep learning techniques excel in many aspects, such as mining non-linear aspects of interactions and handling sparse data [22–24]. Therefore, some scholars have attempted to incorporate neural networks into cognitive diagnosis tasks to improve the model fitting ability. The study [25] proposed the DIRT model, which implements a neural network-based IRT model that can utilize textual information from examination questions in the cognitive diagnosis process. Study [26] proposed the neural CD model, which combines neural networks with the principle of educational monotonicity and achieves excellent performance on various data sets with good portability. However, in the combination of cognitive diagnostics with neural networks, a large amount of research work has focused more on improving the predictive ability of students’ correct answers without deeply exploring the intermediate products of CD (e.g., students’ skill mastery status) to achieve CD functionality. Therefore, the existing deep cognitive diagnostics, although improving the model fitting ability, cannot better model students’ cognitive processes and lack good explanatory power.
Based on the above research, this paper proposes a two-level enhancement model that integrates deep learning and cognitive diagnosis to diagnostic students’ skill mastery states. On the one hand, the advantages of CDM are used to enhance the interpretability of the models. On the other hand, the unique technical advantages of the deep learning method itself can mine the non-linear relationship between students’ skill states and test questions, and to a certain extent can solve the problem that sparse data is difficult to be used by the model. Based on the combination of the both, not only the interpretability of the model can be improved, but also the accuracy of the model can be improved by mining the non-linear interaction data between students and test questions.
The main contributions of this paper are as follows:
1) We propose an innovative cognitive diagnostic framework based on deep learning for uncovering students’ potential skill mastery status. The model is divided into three modules: enhanced input module, deep training module, and prediction module.
2) By modelling the interaction between student-skills (i.e., student skill mastery) and questions-skills (i.e., Q matrix), we obtain students’ question proficiency and further apply filtering network mechanisms to filter noise, such as guessing and slipping, to enhance the interpretability of the model.
3) The neural network is further used to simulate the cognitive process of the students’ responses to the questions, and to train the students skill mastery by predicting the results of their responses. The method uses multiple neurons to model the non-linear interaction between students and questions, improving the fitting ability of traditional models.
In this section, we discuss the research related to Traditional CDM and Deep CDM, and compare the two modeling approaches, as shown in Fig. 1. Among them, the Traditional CDM is to model the interaction between students and test questions through manually constructed item response functions; Deep CDN is capable of mining richer information by modeling the complex interactions between students and test questions through neural network.
Traditional CDMs model the interaction between students and test questions by constructing item response functions using probabilistic statistical models through the monotonicity assumption that the more skills a student examines on a test question, the greater the probability of answering the question correctly. IRT is a classic psychometric theory in the field of education and a representative of the continuum-type of CD model. It is assumed that the subject has a potential trait, which is a concept based on the observation and analysis of the subject’s responses, usually referring to potential abilities. In a continuum-type model based on IRT, students’ potential traits are generally modeled along a continuum of values. IRT proposed the earliest Rasch model [9]. With consideration of how to define both the ability of the subject and the difficulty of the test questions, the item response function is defined follow:
where
CDMs can better model students’ cognitive states from the skill dimension, based on the Q matrix [10]. Among the existing CD models, the DINA model has been most extensively studied. The DINA model is representative of discrete CD models [11]. It treats students’ cognitive state as a binary discrete vector, with two dimensions indicating students’ mastery or non-mastery of the relevant knowledge concepts, and usually assumes an ‘and’ relationship between the mastery of each knowledge concept, the item response function is constructed:
where
The study [27] proposed the definition of Deep Cognitive Diagnostic Model (DCDM), which is the use of deep learning techniques to enhance the cognitive diagnostic capabilities of traditional cognitive diagnostic models. The study [28] applied rough sets and neural networks to psychometric theory. Study [29] used cluster analysis for CD. However, in the integration of cognitive diagnoses with neural networks, many research efforts have focused more on improving the predictive ability of students’ correct answers without an in-depth exploration of the intermediate product of cognitive diagnoses (e.g., students’ skill mastery status). In conclusion, the current deep CDM has made some advancements in simulating non-linear interactions and improving model fit, but it lacks interpretability and does not allow for the diagnosis of student skill mastery.
The difficulty of deep CD is its widely criticized black box nature, i.e., the model parameters are difficult to interpret. The CD needs to get the students’ level of mastery on each knowledge point, so this is the hurdle that must be crossed. Meanwhile, neural networks also present new opportunities, as their powerful fitting abilities allow their CDM to learn more complex and realistic interaction functions from the data. This research proposes a model that integrates deep learning with CD the goal of diagnosing students’ skill states.
The existing DCDM has weak diagnostic ability in the state of students’ skill mastery, and lacks interpretation. In this section, we propose a two-level enhanced deep cognitive diagnosis (EDCD) model based on deep learning for mining students’ skill mastery. The EDCDM framework, as shown in Fig. 2, mainly includes input module, training module and prediction module.
The main task of the initialization module is to randomly initialize the generated parameters for use in the two-level optimization input module. The parameters are taken from the traditional CDMs and represented in the neural network in the form of appropriate data. It is assumed that there are J test questions examining K skills, answered by student I.
Question features: matrix
Student features:
Monotonicity Assumption: The probability of students correctly answering the questions increases monotonously with the student’s skill mastery.
Here the parameters are initialized in the form of a tensor parameter. Initialize the continuous mastery of each skill for each student. Such an initialization pattern provides a wider correction space for the inverse feedback iteration and brings the student’s skill mastery closer to the true value.
This input module obtains students’ question proficiency through a two-layer mechanism: the first layer is the ideal state; the second layer is for filtering noise. The main function of the two-level enhanced input module, shown in Fig. 2 (left), is to combine various randomly generated parameters in the form of DINA formulas by Eq. (2) and then provide the results as input to the next module.
First layer: The first layer obtains the ideal question proficiency state of the students by
Second layer: A Slipping gate (
The parameter
The core of EDCDM is to use the parameter updating mechanism of neural networks to bring the involved parameters closer to the true values by adjusting the mapping relationship between inputs and outputs. Then the continuous predicted values of potential features (such as the student’ skill mastery) are obtained, and the predicted values gradually approach the potential true values, and the prediction accuracy gradually improves.
3.3.1 Neural Network Structure
This section combines the information of two dimensions, test question difficulty and student mastery, and after receiving the mixed input
We use a splicing technique between the second linear-sigmoid layer and the third linear-sigmoid layer to splice the mapping product
This paper uses a method to calculate the error gradient under each weight in real time. It is a classical method for training neural networks in combination with optimization methods (such as gradient descent, etc.) and consists of two parts: incentive propagation and weight update. In the incentive propagation phase, each iteration is performed in two steps:
1) Input the training results into the network to obtain the excitation response;
2) Differentiate the excitation response from the corresponding output target to obtain the response error of the output and hidden layers.
In the weight update phase, two steps are performed for each weight:
1) Multiply the input excitation and response errors to obtain the gradient of the weights;
2) Use this gradient to multiply the learning rate, then take its inverse and add it to the weights.
In this model, the parameter update mechanism plays the role of updating the parameters for fitting, as described by the following equation.
where,
In the structure of the feedback neural network, this model chooses the Cross Entropy Loss Function (CELF) as the loss function to measure the loss between the predicted and true values, and proves the validity of the model by pursuing a lower loss value. The CELF equation can be defined as Eq. (9).
The deep diagnosis module builds on the first two by using a good fit of the neural network to predict the probability of the correct answer of the student by the Eqs. (8) and (9). In this process, rich intermediate results can be obtained due to the rational use of inverse feedback. For example, the degree of the student’s mastery of a specific skill, the difficulty of the knowledge point in the exam test, etc. Such intermediate findings are a useful complement to cognitive diagnosis.
In this section, we conduct several experiments to demonstrate the effectiveness of the EDCDM framework from multiple perspectives, using multiple datasets and comparing them to the baseline model.
We used three real-world datasets in our experiments, namely PISA 2015 [4], Math [5], and Assist [6]. Table 1 summarizes the information from these datasets.
Math: The Math dataset is a dataset of objective and subjective responses to a final math exam for high school students that contains subjective questions with continuous response scores.
PISA 20151: PISA is the definitive global online test with high quality items. In this study, 17 computer-scaled dichotomous mathematics test items were selected for analysis.
Assist2: Assist is an open dataset, Assistment 2009–2010 “skill builder”, which provides only student response logs and questions corresponding to knowledge concepts. Student response data in this dataset is extremely sparse.
In order to assess the validity of the model, classical as well as the latest methods were selected as the baseline model. The contrasting models are described as follows.
IRT: IRT is one of the most classical models for cognitive diagnosis. It is modeled by attribute values between students and test questions, and the diagnostic result of IRT is a comprehensive but vague ability value.
MIRT: Multidimensional Item Response Theory, an item response theory with a multidimensional CDM, is an extension of the one-dimensional IRT.
DINA: DINA model enables micro-diagnosis of students’ mastery of specific knowledge concepts through the Q matrix.
Neural CDM: Neural CDM is a deep learning-based cognitive diagnostic model that incorporates item response theory to diagnose the cognitive properties of students and questions.
In this paper, we use three widely used metrics in the domain to evaluate the predictive performance of all models. From a regression perspective, we chose Root Mean Square Error (RMSE) to quantify the distance between predicted and actual scores. The smaller the value, the better the result. In addition, we consider the prediction problem as a classification task, where a response record with a score of 1 (0) indicates a correct (wrong) answer. Therefore, we use two metrics, namely, prediction accuracy (ACC), and area under the ROC curve (AUC), for the measurement. In general, an AUC or ACC value of 0.5 represents the performance prediction result of a random guess, and the larger the value, the better of the prediction ability of model.
This experiment is carried out in the following areas to confirm the performance of the suggested model: (1) evaluation of Baselines model prediction performance on various data sets; (2) evaluation of ablation experiments with various components; (3) confirmation of model monotonicity; and (4) visualization of diagnostic outcomes.
4.3.1 Student Performance Prediction
To evaluate whether the prediction performance of the model is generalizable across different datasets, we divided each dataset into an 80% training set and a 20% test set. The experimental results are shown in Table 2. The EDCD model outperformed almost all the baselines on both datasets, proving the effectiveness of the proposed model. For the MATH dataset, which contains both objective and subjective questions, the subjective part was discarded because the IRT, MIRT, and DINA models could not be applied to the subjective questions. For the PISA dataset, EDCDM also outperformed all baseline models in terms of prediction results. For the relatively sparse Assist dataset, the EDCDM model also produced the best experimental results overall. The interpretability of the traditional model was carried over to the EDCDM model, which also performed better overall than the neural CDM. The experiments showed that the traditional cognitive diagnostic model performed poorly on the sparse dataset.
The learning curve can also be utilized as a foundation for judgment to confirm that the suggested model has achieved convergence. To be more precise, 20% of the dataset is used as the test set, while the remaining 80% is used as the training set. We used three real-world datasets in our experiments, namely PISA 2015, mathematics, and auxiliary. During the training period, we record train loss, test loss and AUC metrics. Finally, EDCDM depicts the learning curve for each dataset, as shown in Fig. 3. The training loss, test loss, and AUC metric curves on all three datasets eventually level off, as shown in the learning curve plot above. Therefore, it can be assumed that given the relative number of current datasets, individual metrics can be made to converge and produce the desired training effect.
To verify the effect of each component in the proposed model on the accuracy of the EDCDM, we chose PISA and Math two datasets for the ablation experiments. We extended the EDCDM to verify the validity of the method used in the model with a 20% test rate. Table 3 gives the results of each index of the three extended experiments No_diff, No_skill, and No_gate on each dataset. No_diff is to disregard the difficulty of the test questions and set the difficulty parameter to 0; No_skill is to disregard the Q matrix, i.e., the skills tested by the test questions, similar to MIRT, which considers students to have multidimensional abilities; No_ gate is to disregard the factors of Slipping and Guessing, and to consider that students can answer the test questions correctly by mastering the skills tested by the test questions, without the mechanism of error tolerance.
Compared to No_diff, the accuracy metric ACC metric improved by 0.019 on average and the error metric RMSE improved by 0.013 for EDCDM, indicating the validity of the difficulty representation of the test questions.
Compared to No_skill, the accuracy metric ACC metric improved by 0.026 on average and the error metric RMSE improved by 0.013 for EDCDM, indicating a positive effect of the skill interaction component on the prediction task.
Compared to No_diff, the accuracy metric ACC improved by 0.04 on average and the error metric RMSE improved by 0.03 for EDCDM, indicating that the screening gate mechanism (i.e., slipping and guessing factors) has an important role in the proposed model.
To assess the monotonicity of the EDCDM (i.e., the higher the skill mastery, the higher the probability of answering the test questions correctly), we conducted experiments on different data sets. Fig. 4 shows the correlation between student’s skill mastery and response score, with three subplots representing the PISA, MATH, ASSIST datasets respectively.
As can be seen from the figures, both subplots Figs. 4a and 4b perfectly fit the monotonicity assumption of the cognitive diagnostic theory. indicating that the proposed model performs well on the traditional dense data set. Fig. 4c is also consistent with the monotonicity assumption overall, however there are some outliers around a skill mastery level of 0.6. The students in these outliers have extremely few answer records, according to subsequent study. This finding shows that the monotonicity of the model needs to be further tuned on sparse dataset.
We assessed the interpretability of the EDCDM by showing the distribution of students’ skill states through box plots and kernel density function plots. As shown in Fig. 5a for the neural CDM based on the mathematical dataset, the heat map of students’ mastery of all skills is almost the same color, indicating that there are almost no data fluctuations in their simulated values. In addition, the students’ mastery status for all attributes is mainly concentrated around 0.5 without any significant outliers. As shown in Fig. 5b, the EDCDM model based on the mathematical dataset has a more reasonable distribution of students’ mastery of all qualities, with balanced fluctuations, and has a more practical reference value.
Fig. 6 shows the distribution of EDCDM mastery for each skill in the PISA and MATH datasets respectively through box plots and kernel density distribution plots. Fig. 6a shows the distribution of mastery for each skill in the MATH dataset, with a median mastery of about 0.6 for each skill. Skill 2 is below average, indicating that it is more difficult. Skill 1 and Skill 3 were mastered at a much higher level than the others, indicating that they were less difficult. In addition, the distribution of students’ mastery of these two skills is spread out, suggesting that they can be used to more effectively differentiate students with different skills. Fig. 6b shows the distribution of mastery for each skill in the PISA dataset, with a median mastery of about 0.6 for each skill, and a median of about 0.7 for skill 4, indicating that this skill is relatively easy for students.
To address the lack of modeling complex interactions between students and test questions in existing cognitive diagnostic models, this paper proposes a two-level augmented cognitive diagnostic model based on deep learning. EDCDM models students’ mastery of test questions through cognitive factors and then trains nonlinear interactions between students and test questions through neural networks to diagnose students’ skill mastery. By comparing with several baseline models, the following conclusions were obtained: (1) compared with other baseline models, EDCDM has higher accuracy in predicting student performance; (2) compared with other deep cognitive diagnostic models, EDCDM has good discrimination in diagnosing learners’ cognitive states, which makes the model have better explanatory power. With the continuous development and improvement of online learning, the research on cognitive diagnostic models and their applications for personalized learning will likely achieve technical breakthroughs in many aspects. Future work can be carried out in two aspects, namely dynamic cognitive modeling of students and cross-disciplinary knowledge transfer learning analysis, in order to realize the modeling of students’ cross-disciplinary cognitive states, which is important for a comprehensive understanding of students’ learning states and more accurate personalized learning tutoring.
Funding Statement: The authors received no specific funding for this study.
Availability of Data and Materials: The experimental data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding the present study.
1https://www.oecd.org/pisa/data/2015database/.
2https://sites.google.com/site/assistmentsdata/home/assistment-2009-2010-data/skill-builder-data-2009-2010.
References
1. L. Tetzlaff, F. Schmiedek and G. Brod, “Developing personalized education: A dynamic framework,” Educational Psychology Review, vol. 33, pp. 863–882, 2021. [Google Scholar]
2. A. Sun and X. Chen, “Online education and its effective practice: A research review,” Journal of Information Technology Education: Research, vol. 15, 2016. [Google Scholar]
3. Y. Cui, M. Gierl and Q. Guo, “Statistical classification for cognitive diagnostic assessment: An artificial neural network approach,” Educational Psychology, vol. 36, no. 6, pp. 1065–1082, 2016. [Google Scholar]
4. A. Peña-Ayala, “Educational data mining: A survey and a data mining-based analysis of recent works,” Expert Systems with Applications, vol. 41, no. 4, pp. 1432–1462, 2014. [Google Scholar]
5. Y. Zhang, Y. Yun, R. An, J. Cui, H. Dai et al., “Educational data mining techniques for student performance prediction: Method review and comparison analysis,” Frontiers in Psychology, vol. 12, pp. 698490, 2021. [Google Scholar] [PubMed]
6. C. Liu and Y. Cheng, “An application of the support vector machine for attribute-by-attribute classification in cognitive diagnosis,” Applied Psychological Measurement, vol. 42, no. 1, pp. 58–72, 2018. [Google Scholar] [PubMed]
7. R. S. Baker, T. Martin and L. M. Rossi, “Educational data mining and learning analytics,” in The Wiley Handbook of Cognition and Assessment: Frameworks, Methodologies, and Applications, pp. 379–396, 2016. [Google Scholar]
8. C. F. Lin, Y. -C Yeh, Y. H. Hung and R. I. Chang, “Data mining for providing a personalized learning path in creativity: An application of decision trees,” Computers & Education, vol. 68, pp. 199–210, 2013. [Google Scholar]
9. N. A. Samah, N. Yahaya and M. B. Ali, “Individual differences in online personalized learning environment,” Educational Research and Reviews, vol. 6, no. 7, pp. 516–521, 2011. [Google Scholar]
10. K. S. McCarthy, M. Watanabe, J. Dai and D. S. McNamara, “Personalized learning in iSTART: Past modifications and future design,” Journal of Research on Technology in Education, vol. 52, no. 3, pp. 301–321, 2020. [Google Scholar]
11. S. Y. Chen, P. -R. Huang, Y. -C. Shih and L. -P. Chang, “Investigation of multiple human factors in personalized learning,” Interactive Learning Environments, vol. 24, no. 1, pp. 119–141, 2016. [Google Scholar]
12. A. B. F. Mansur, N. Yusof and A. H. Basori, “Personalized learning model based on deep learning algorithm for student behaviour analytic,” Procedia Computer Science, vol. 163, pp. 125–133, 2019. [Google Scholar]
13. C. Romero and S. Ventura, “Educational data mining: A review of the state of the art,” IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews), vol. 40, no. 6, pp. 601–618, 2010. [Google Scholar]
14. J. de la Torre and J. A. Douglas, “Higher-order latent trait models for cognitive diagnosis,” Psychometrika, vol. 69, no. 3, pp. 333–353, 2004. [Google Scholar]
15. J. de la Torre, “DINA model and parameter estimation: A didactic,” Journal of Educational and Behavioral Statistics, vol. 34, no. 1, pp. 115–130, 2008. [Google Scholar]
16. J. de la Torre, “The generalized DINA model framework,” Psychometrika, vol. 76, no. 2, pp. 179–199, 2011. [Google Scholar]
17. M. von Davier, “The log-linear cognitive diagnostic model (LCDM) as a special case of the general diagnostic model (GDM),” ETS Research Report Series, vol. 2014, no. 2, pp. 1–13, 2014. [Google Scholar]
18. A. C. George and A. Robitzsch, “Validating theoretical assumptions about reading with cognitive diagnosis models,” International Journal of Testing, vol. 21, no. 2, pp. 105–129, 2021. [Google Scholar]
19. W. Song, L. Wen, L. Gao and X. Li, “Unsupervised fault diagnosis method based on iterative multi-manifold spectral clustering,” IET Collaborative Intelligent Manufacturing, vol. 1, no. 2, pp. 48–55, 2019. [Google Scholar]
20. M. Xi, “A study on the generalized cognitive diagnosis model of students based on BN network,” Modern Electric Electronic Technology, vol. 41, no. 24, pp. –79-81, 2018. [Google Scholar]
21. Y. Wang, “A cognitive diagnostic model of students’ psychological multilevel rating based on HO-DINA model research,” Modern Electric Electronic Technology, vol. 41, no. 2, pp. 53–55, 2018. [Google Scholar]
22. Q. Liu, R. Wu, E. Chen, G. Xu, Y. Su et al., “Fuzzy cognitive diagnosis for modelling examinee performance,” ACM Transactions on Intelligent Systems and Technology, vol. 9, no. 4, pp. 1–26, 2018. [Google Scholar]
23. Y. Zhou, Q. Liu, J. Wu et al., “Modeling context-aware features for cognitive diagnosis in student learning,” in Proceedings of the 27th ACM SIGKDD conference on knowledge discovery & data mining, 2021, pp. 2420–2428. [Google Scholar]
24. J. Li, F. Wang, Q. Liu, M. Zhu, W. Huang et al., “Hiercdf: A bayesian network-based hierarchical cognitive diagnosis framework,” in Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 904–913, 2022. [Google Scholar]
25. S. Cheng, Q. Liu, E. Chen, Z. Huang, Y. Chen et al., “Dirt: Deep learning enhanced item response theory for cognitive diagnosis,” in Proceedings of the 28th ACM International Conference on Information and Knowledge Management, pp. 2397–2400, 2019. [Google Scholar]
26. F. Wang, Q. Liu, E. Chen, Z. Huang, Y. Chen et al., “Neural cognitive diagnosis for intelligent education systems,” Proceedings of the AAAI Conference on Artificial Intelligence., vol. 34, no. 4, pp. 6153–6161, 2020. [Google Scholar]
27. L. Gao, Z. Zhao, C. Li, J. Zhao, Q. Zeng et al., “Deep cognitive diagnosis model for predicting students’ performance,” Future Generation Computer Systems, vol. 126, pp. 252–262, 2022. [Google Scholar]
28. H. Yang, T. Qi, J. Li, L. Guo, M. Ren et al., “A novel quantitative relationship neural network for explainable cognitive diagnosis model,” Knowledge-Based Systems, vol. 250, pp. 109156, 2022. [Google Scholar]
29. C. Y. Chiu, J. A. Douglas and X. Li, “Cluster analysis for cognitive diagnosis: Theory and applications,” Psychometrika, vol. 74, pp. 633–665, 2009. [Google Scholar]
Cite This Article
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.