Arrow Research search

Author name cluster

Ziwen Wang

Possible papers associated with this exact author name in Arrow. This page groups case-insensitive exact name matches and is not a full identity disambiguation profile.

10 papers
1 author row

Possible papers

10

AAAI Conference 2026 Conference Paper

Debiased Cognitive Diagnosis: A Contrastive Counterfactual Modeling Method via Variational Autoencoder

  • Shangshang Yang
  • Xuewen Duan
  • Xiaoshan Yu
  • Ziwen Wang
  • Haiping Ma
  • Xingyi Zhang

Cognitive diagnosis (CD), inferring student knowledge mastery based on historical response records, is crucial for personalized educational services such as adaptive practice and learning path planning. Existing CD models were built based on the assumption that student's response data is integral, overlooking the nonrandom missingness of data caused by student answering exercises selectively. This missingness generally leads to biased and incomplete observations, where confounders, such as selection bias and exposure bias, significantly undermine the accuracy of student knowledge modeling. To address missingness, we propose a Debiased Cognitive Diagnosis (DBCD) framework through the perspective of counterfactual modeling to remove exogenous confounders from the response data. Specifically, the proposed DBCD achieves debiasing for CD by applying the idea of contrastive learning to constrain the model's prediction distributions on both factual and counterfactual data. For a student, the factual data is his/her original response records, while the counterfactual data is generated by sampling the same number of exercises from all exercises of each concept through a similarity-based counterfactual sampling strategy. Considering the difficulty of directly removing the exogenous confounders for student, we devise a β-Variational Autoencoder to model their exogenous confounders within the latent representations of knowledge proficiency by leveraging exercise priors and student response patterns. Then, the learned representations are further combined with the vanilla student's ability embedding via a gating mechanism-based fusion for final diagnosis prediction of the model. Extensive experiments on real-world educational datasets demonstrate that the proposed DBCD effectively mitigates confounders and even outperforms existing methods, thereby validating the feasibility and effectiveness of the DBCD framework.

JBHI Journal 2026 Journal Article

Instance-Based Transfer Learning With Similarity-Aware Subject Selection for Cross-Subject SSVEP-Based BCIs

  • Ziwen Wang
  • Yue Zhang
  • Zhiqiang Zhang
  • Sheng Quan Xie
  • Alexander Lanzon
  • William P. Heath
  • Zhenhong Li

Steady-state visual evoked potential (SSVEP)-based brain-computer interfaces (BCIs) can achieve high recognition accuracy with sufficient training data. Transfer learning presents a promising solution to alleviate data requirements for the target subject by leveraging data from source subjects; however, effectively addressing individual variability among both target and source subjects remains a challenge. This paper proposes a novel transfer learning framework, termed instance-based task-related component analysis (iTRCA), which leverages knowledge from source subjects while considering their individual contributions. iTRCA extracts two types of features: (1) the subject-general feature, capturing shared information between source and target subjects in a common latent space, and (2) the subject-specific feature, preserving the unique characteristics of the target subject. To mitigate negative transfer, we further design an enhanced framework, subject selection-based iTRCA (SS-iTRCA), which integrates a similarity-based subject selection strategy to identify appropriate source subjects for transfer based on their task-related components (TRCs). Comparative evaluations on the Benchmark, BETA, and a self-collected dataset demonstrate the effectiveness of the proposed iTRCA and SS-iTRCA frameworks. This study provides a potential solution for developing high-performance SSVEP-based BCIs with reduced target subject data.

AAAI Conference 2026 Conference Paper

PEOAT: Personalization-Guided Evolutionary Question Assembly for One-Shot Adaptive Testing

  • Xiaoshan Yu
  • Ziwei Huang
  • Shangshang Yang
  • Ziwen Wang
  • Haiping Ma
  • Xingyi Zhang

With the rapid advancement of intelligent education, Computerized Adaptive Testing (CAT) has attracted increasing attention by integrating educational psychology with deep learning technologies. Unlike traditional paper-and-pencil testing, CAT aims to efficiently and accurately assess ex- aminee abilities by adaptively selecting the most suitable items during the assessment process. However, its real-time and sequential nature presents limitations in practical scenarios, particularly in large-scale assessments where interaction costs are high, or in sensitive domains such as psychological evaluations where minimizing noise and interfer- ence is essential. These challenges constrain the applicability of conventional CAT methods in time-sensitive or resource- constrained environments. To this end, we first introduce a novel task called one-shot adaptive testing (OAT), which aims to select a fixed set of optimal items for each test-taker in a one-time selection. Meanwhile, we propose PEOAT, a Personalization-guided Evolutionary question assembly framework for One-hot Adaptive Testing from the perspec- tive of combinatorial optimization. Specifically, we began by designing a personalization-aware initialization strategy that integrates differences between examinee ability and ex- ercise difficulty, using multi-strategy sampling to construct a diverse and informative initial population. Building on this, we proposed a cognitive-enhanced evolutionary framework incorporating schema-preserving crossover and cognitively guided mutation to enable efficient exploration through infor- mative signals. To maintain diversity without compromising fitness, we further introduced a diversity-aware environmen- tal selection mechanism. The effectiveness of PEOAT is val- idated through extensive experiments on two datasets, com- plemented by case studies that uncovered valuable insights.

TIST Journal 2026 Journal Article

Reconciling Cognitive Modeling with Knowledge Forgetting: A Continuous Time-aware General Neural Network Framework

  • Ziwen Wang
  • Jingyuan Wang
  • Haiping Ma
  • Hengshu Zhu
  • Shangshang Yang
  • Xiaoshan Yu
  • Shuhuan Liu
  • Haifeng Zhang

Cognitive modeling, as an emerging technology in the field of computer-aided education, aims to explore students’ knowledge levels and learning abilities to achieve various intelligent educational applications. Although some existing work focuses on addressing the problem of student forgetting, it is still a less explored area how to naturally integrate the forgetting effect caused by the time interval between answering exercises into student knowledge state modeling. Additionally, traditional cognitive modeling methods mostly assume that students answer exercises one by one, which often does not align with real answering behavior and cannot be directly extended to diverse learning scenarios. Therefore, in this article, we propose a Continuous Time-based Neural Cognitive (CT-NC) framework and several implemented models (CT-NCM and two extensions) to effectively integrate the dynamic and continuous characteristics of knowledge forgetting into student learning process modeling, making it more natural. Specifically, we adopt a specially designed learning event encoding method to adjust the neural Hawkes process to capture the relationship between knowledge learning and forgetting over continuous time. Furthermore, we propose a customizable learning function to jointly model the changes in different knowledge states and their interaction with each practice moment. In the end, we demonstrate an extension CT-NCM+ that can adapt well to diverse learning scenarios, indicating that CT-NCM can solve real-world problems by flexibly adjusting its structure. Extensive experimental results on real datasets clearly demonstrate that CT-NCM and CT-NCM+ outperform the current state-of-the-art KT methods in student performance prediction, while our work points out a realistic research direction for KT and demonstrates its interpretability in knowledge learning visualization.

IJCAI Conference 2025 Conference Paper

Endowing Interpretability for Neural Cognitive Diagnosis by Efficient Kolmogorov-Arnold Networks

  • Shangshang Yang
  • Linrui Qin
  • Xiaoshan Yu
  • Ziwen Wang
  • Xueming Yan
  • Haiping Ma
  • Ye Tian

Cognitive diagnosis is crucial for intelligent education because of its ability to reveal students' proficiency in knowledge concepts. Although neural network-based neural cognitive diagnosis models (CDMs) have exhibited significantly better performance than traditional models, neural cognitive diagnosis is criticized for the poor model interpretability due to the multi-layer perceptron(MLP) employed, even with the monotonicity assumption. Therefore, this paper proposes to empower the interpretability of neural cognitive diagnosis models through efficient Kolmogorov-Arnold networks (KANs), named KAN2CD, where KANs are used to enhance interpretability in two manners. Specifically, in the first manner, KANs are directly used to replace the used MLPs in existing neural CDMs; while in the second manner, the student embedding, exercise embedding, and concept embedding are directly processed by several KANs, and then their outputs are further combined and learned in a unified KAN to get final predictions. Besides, the implementation of original KANs is modified without affecting the interpretability to overcome the problem of training KANs slowly. Extensive experiments show KAN2CD outperforms traditional CDMs and slightly surpasses existing neural CDMs, and its learned structures ensure interpretability on par with traditional CDMs and better than neural CDMs. The datasets, associated code, and more experimental results are available at https: //github. com/null233QAQ/KAN2CD.

AAAI Conference 2025 Conference Paper

Explicit and Implicit Examinee-Question Relation Exploiting for Efficient Computerized Adaptive Testing

  • Changqian Wang
  • Shangshang Yang
  • Siyu Song
  • Ziwen Wang
  • Haiping Ma
  • Xingyi Zhang
  • Bo Jin

Computerized adaptive testing(CAT) is a crucial task in computer-aided education, which aims to adaptively select suitable question to diagnose examinees' ability status. Existing CAT approaches enhance selection performance by exploring examinee-question(E-Q) relation. These approaches either exclusively utilize explicit E-Q relation. For instance, policy-based approaches determine question selection based on predefined criteria. While effective in adapting to changes in question banks, these methods often entail significant computational costs in searching for suitable questions. Conversely, some studies focus solely on implicit E-Q relation. For example, learning-based approaches train agents to efficiently select questions by learning from large-scale datasets. However, they may struggle with newly introduced questions. Additionally, most of these existing question selectors are based on greedy strategies, which potentially overlooks promising quuestions. To bridge the above two types of approaches, we propose a novel framework named Relation Exploiting-based CAT(RECAT) by exploring and exploiting the implicit and explicit examinee-question relation. Specifically, we first define an examinee true ability-oriented selection objective to select more suitable questions. Then, to learn the implicit E-Q relation, we design a question selector, which explores the examinee ability and generates best-fitting questions for specific examinee ability from two aspects, including generation consistency and knowledge matching. The former aims to maximize the likelihood estimation of the implicit E-Q relation learning process, while the latter is employed to fit the distribution of real questions. To fully exploit explicit E-Q relation, we generate a high-quality candidate set for the given examinee's ability using implicit E-Q relation, which streamlines the search process, minimizing selection latency. We demonstrate the effectiveness and efficiency of our framework through comprehensive experiments on real-world datasets.

NeurIPS Conference 2025 Conference Paper

Knowledge Starts with Practice: Knowledge-Aware Exercise Generative Recommendation with Adaptive Multi-Agent Cooperation

  • Yangtao Zhou
  • Hua Chu
  • Chen Chen
  • Ziwen Wang
  • Jiacheng Liu
  • Jianan Li
  • Yueying Feng
  • Xiangming Li

Adaptive learning, which requires the in-depth understanding of students' learning processes and rational planning of learning resources, plays a crucial role in intelligent education. However, how to effectively model these two processes and seamlessly integrate them poses significant implementation challenges for adaptive learning. As core learning resources, exercises have the potential to diagnose students' knowledge states during the learning processes and provide personalized learning recommendations to strengthen students' knowledge, thereby serving as a bridge to boost student-oriented adaptive learning. Therefore, we introduce a novel task called Knowledge-aware Exercise Generative Recommendation (KEGR). It aims to dynamically infer students' knowledge states from their past exercise responses and customizably generate new exercises. To achieve KEGR, we propose an adaptive multi-agent cooperation framework, called ExeGen, inspired by the excellent reasoning and generative capabilities of LLM-based AI agents. Specifically, ExeGen coordinates four specialized agents for supervision, knowledge state perception, exercise generation, and quality refinement through an adaptive loop workflow pipeline. More importantly, we devise two enhancement mechanisms in ExeGen: 1) A human-simulated knowledge perception mechanism mimics students' cognitive processes and generates interpretable knowledge state descriptions via demonstration-based In-Context Learning (ICL). In this mechanism, a dual-matching strategy is further designed to retrieve highly relevant demonstrations for reliable ICL reasoning. 2) An exercise generation-adversarial mechanism collaboratively refines exercise generation leveraging a group of quality evaluation expert agents via iterative adversarial feedback. Finally, a comprehensive evaluation protocol is carefully designed to assess ExeGen. Extensive experiments on real-world educational datasets and a practical deployment in college education demonstrate the effectiveness and superiority of ExeGen. The code is available at https: //github. com/dsz532/exeGen.

JMLR Journal 2025 Journal Article

Randomly Projected Convex Clustering Model: Motivation, Realization, and Cluster Recovery Guarantees

  • Ziwen Wang
  • Yancheng Yuan
  • Jiaming Ma
  • Tieyong Zeng
  • Defeng Sun

In this paper, we propose a randomly projected convex clustering model for clustering a collection of $n$ high dimensional data points in $\mathbb{R}^d$ with $K$ hidden clusters. Compared to the convex clustering model for clustering original data with dimension $d$, we prove that, under some mild conditions, the perfect recovery of the cluster membership assignments of the convex clustering model, if exists, can be preserved by the randomly projected convex clustering model with embedding dimension $m = O(\epsilon^{-2}\log(n))$, where $\epsilon > 0$ is some given parameter. We further prove that the embedding dimension can be improved to be $O(\epsilon^{-2}\log(K))$, which is independent of the number of data points. We also establish the recovery guarantees of our proposed model with uniform weights for clustering a mixture of spherical Gaussians. Extensive numerical results demonstrate the robustness and superior performance of the randomly projected convex clustering model. The numerical results will also demonstrate that the randomly projected convex clustering model can outperform other popular clustering models on the dimension-reduced data, including the randomly projected K-means model. [abs] [ pdf ][ bib ] &copy JMLR 2025. ( edit, beta )

NeurIPS Conference 2025 Conference Paper

Variational Supervised Contrastive Learning

  • Ziwen Wang
  • Jiajun Fan
  • Thao Nguyen
  • Heng Ji
  • Ge Liu

Contrastive learning has proven to be highly efficient and adaptable in shaping representation spaces across diverse modalities by pulling similar samples together and pushing dissimilar ones apart. However, two key limitations persist: (1) Without explicit regulation of the embedding distribution, semantically related instances can inadvertently be pushed apart unless complementary signals guide pair selection, and (2) excessive reliance on large in-batch negatives and tailored augmentations hinders generalization. To address these limitations, we propose Variational Supervised Contrastive Learning (VarCon), which reformulates supervised contrastive learning as variational inference over latent class variables and maximizes a posterior-weighted evidence lower bound (ELBO) that replaces exhaustive pair-wise comparisons for efficient class-aware matching and grants fine-grained control over intra-class dispersion in the embedding space. Trained exclusively on image data, our experiments on CIFAR-10, CIFAR-100, ImageNet-100, and ImageNet-1K show that VarCon (1) achieves state-of-the-art performance for contrastive learning frameworks, reaching 79. 36% Top-1 accuracy on ImageNet-1K and 78. 29% on CIFAR-100 with a ResNet-50 encoder while converging in just 200 epochs; (2) yields substantially clearer decision boundaries and semantic organization in the embedding space, as evidenced by KNN classification, hierarchical clustering results, and transfer-learning assessments; and (3) demonstrates superior performance in few-shot learning than supervised baseline and superior robustness across various augmentation strategies.

NeurIPS Conference 2024 Conference Paper

DisenGCD: A Meta Multigraph-assisted Disentangled Graph Learning Framework for Cognitive Diagnosis

  • Shangshang Yang
  • Mingyang Chen
  • Ziwen Wang
  • Xiaoshan Yu
  • Panpan Zhang
  • Haiping Ma
  • Xingyi Zhang

Existing graph learning-based cognitive diagnosis (CD) methods have made relatively good results, but their student, exercise, and concept representations are learned and exchanged in an implicit unified graph, which makes the interaction-agnostic exercise and concept representations be learned poorly, failing to provide high robustness against noise in students' interactions. Besides, lower-order exercise latent representations obtained in shallow layers are not well explored when learning the student representation. To tackle the issues, this paper suggests a meta multigraph-assisted disentangled graph learning framework for CD (DisenGCD), which learns three types of representations on three disentangled graphs: student-exercise-concept interaction, exercise-concept relation, and concept dependency graphs, respectively. Specifically, the latter two graphs are first disentangled from the interaction graph. Then, the student representation is learned from the interaction graph by a devised meta multigraph learning module; multiple learnable propagation paths in this module enable current student latent representation to access lower-order exercise latent representations, which can lead to more effective nad robust student representations learned; the exercise and concept representations are learned on the relation and dependency graphs by graph attention modules. Finally, a novel diagnostic function is devised to handle three disentangled representations for prediction. Experiments show better performance and robustness of DisenGCD than state-of-the-art CD methods and demonstrate the effectiveness of the disentangled learning framework and meta multigraph module. The source code is available at https: //github. com/BIMK/Intelligent-Education/tree/main/DisenGCD.