Arrow Research search

Author name cluster

Ziqi Chen

Possible papers associated with this exact author name in Arrow. This page groups case-insensitive exact name matches and is not a full identity disambiguation profile.

5 papers
1 author row

Possible papers

5

AAAI Conference 2025 Conference Paper

Conditional Diffusion Models Based Conditional Independence Testing

  • Yanfeng Yang
  • Shuai Li
  • Yingjie Zhang
  • Zhuoran Sun
  • Hai Shu
  • Ziqi Chen
  • Renming Zhang

Conditional independence (CI) testing is a fundamental task in modern statistics and machine learning. The conditional randomization test (CRT) was recently introduced to test whether two random variables, X and Y, are conditionally independent given a potentially high-dimensional set of random variables, Z. The CRT operates exceptionally well under the assumption that the conditional distribution X|Z is known. However, since this distribution is typically unknown in practice, accurately approximating it becomes crucial. In this paper, we propose using conditional diffusion models (CDMs) to learn the distribution of X|Z. Theoretically and empirically, it is shown that CDMs closely approximate the true conditional distribution. Furthermore, CDMs offer a more accurate approximation of X|Z compared to GANs, potentially leading to a CRT that performs better than those based on GANs. To accommodate complex dependency structures, we utilize a computationally efficient classifier-based conditional mutual information (CMI) estimator as our test statistic. The proposed testing procedure performs effectively without requiring assumptions about specific distribution forms or feature dependencies, and is capable of handling mixed-type conditioning sets that include both continuous and discrete variables. Theoretical analysis shows that our proposed test achieves a valid control of the type I error. A series of experiments on synthetic data demonstrates that our new test effectively controls both type-I and type-II errors, even in high dimensional scenarios.

JBHI Journal 2024 Journal Article

Channel Adaptive and Sparsity Personalized Federated Learning for Privacy Protection in Smart Healthcare Systems

  • Ziqi Chen
  • Jun Du
  • Xiangwang Hou
  • Keping Yu
  • Jintao Wang
  • Zhu Han

With the booming development of Smart Healthcare Systems (SHSs), employing federated learning (FL) in SHS devices has become a research hotspot. FL, as a distributed learning framework, can train models without sharing the original data among users, and then protect the user privacy. Existing research has proposed many methods to improve the security and efficiency of FL, which may not fully consider the characteristics of SHSs. Specifically, the requirements of privacy protection and efficiency pose significant challenges to FL. Current studies have struggled to balance privacy security and efficiency, and the degradation of model training efficiency in SHSs can be critical to patient health. Therefore, to improve the privacy protection of healthcare data and ensure communication efficiency, this work proposes a novel personalized FL framework based on Communication quality and Adaptive Sparsification (pFedCAS). In order to achieve privacy protection, a control unit is proposed and introduced to adjust the sparsity of the local model adaptively. To further improve the training efficiency, a selection unit is added during global model aggregation to select suitable clients for parameter updates. Finally, we validate the proposed method operated on the HAM10000 dataset. Simulation results validate that pFedCAS can not only improve privacy protection, but also gain an improvement of 15% in training accuracy and a reduction of 30% in training costs based on communication quality. The simulation results also validate the excellent robustness of pFedCAS to non-iid data.

NeurIPS Conference 2023 Conference Paper

K-Nearest-Neighbor Local Sampling Based Conditional Independence Testing

  • Shuai Li
  • Yingjie Zhang
  • Hongtu Zhu
  • Christina Wang
  • Hai Shu
  • Ziqi Chen
  • Zhuoran Sun
  • Yanfeng Yang

Conditional independence (CI) testing is a fundamental task in statistics and machine learning, but its effectiveness is hindered by the challenges posed by high-dimensional conditioning variables and limited data samples. This article introduces a novel testing approach to address these challenges and enhance control of the type I error while achieving high power under alternative hypotheses. The proposed approach incorporates a computationally efficient classifier-based conditional mutual information (CMI) estimator, capable of capturing intricate dependence structures among variables. To approximate a distribution encoding the null hypothesis, a $k$-nearest-neighbor local sampling strategy is employed. An important advantage of this approach is its ability to operate without assumptions about distribution forms or feature dependencies. Furthermore, it eliminates the need to derive asymptotic null distributions for the estimated CMI and avoids dataset splitting, making it particularly suitable for small datasets. The method presented in this article demonstrates asymptotic control of the type I error and consistency against all alternative hypotheses. Extensive analyses using both synthetic and real data highlight the computational efficiency of the proposed test. Moreover, it outperforms existing state-of-the-art methods in terms of type I and II errors, even in scenarios with high-dimensional conditioning sets. Additionally, the proposed approach exhibits robustness in the presence of heavy-tailed data.

AAAI Conference 2023 Conference Paper

Nearest-Neighbor Sampling Based Conditional Independence Testing

  • Shuai Li
  • Ziqi Chen
  • Hongtu Zhu
  • Christina Dan Wang
  • Wang Wen

The conditional randomization test (CRT) was recently proposed to test whether two random variables X and Y are conditionally independent given random variables Z. The CRT assumes that the conditional distribution of X given Z is known under the null hypothesis and then it is compared to the distribution of the observed samples of the original data. The aim of this paper is to develop a novel alternative of CRT by using nearest-neighbor sampling without assuming the exact form of the distribution of X given Z. Specifically, we utilize the computationally efficient 1-nearest-neighbor to approximate the conditional distribution that encodes the null hypothesis. Then, theoretically, we show that the distribution of the generated samples is very close to the true conditional distribution in terms of total variation distance. Furthermore, we take the classifier-based conditional mutual information estimator as our test statistic. The test statistic as an empirical fundamental information theoretic quantity is able to well capture the conditional-dependence feature. We show that our proposed test is computationally very fast, while controlling type I and II errors quite well. Finally, we demonstrate the efficiency of our proposed test in both synthetic and real data analyses.