Arrow Research search

Author name cluster

Xinyi Xu

Possible papers associated with this exact author name in Arrow. This page groups case-insensitive exact name matches and is not a full identity disambiguation profile.

21 papers
2 author rows

Possible papers

21

ICLR Conference 2025 Conference Paper

Efficient Top-m Data Values Identification for Data Selection

  • Xiaoqiang Lin
  • Xinyi Xu
  • See-Kiong Ng
  • Bryan Kian Hsiang Low

Data valuation has found many real-world applications, e.g., data pricing and data selection. However, the most adopted approach -- Shapley value (SV) -- is computationally expensive due to the large number of model trainings required. Fortunately, most applications (e.g., data selection) require only knowing the $m$ data points with the highest data values (i.e., top-$m$ data values), which implies the potential for fewer model trainings as exact data values are not required. Existing work formulates top-$m$ Shapley value identification as top-$m$ arms identification in multi-armed bandits (MAB). However, the proposed approach falls short because it does not utilize data features to predict data values, a method that has been shown empirically to be effective. A recent top-$m$ arms identification work does consider the use of arm features while assuming a linear relationship between arm features and rewards, which is often not satisfied in data valuation. To this end, we propose the GPGapE algorithm that uses the Gaussian process to model the \emph{non-linear} mapping from data features to data values, removing the linear assumption. We theoretically analyze the correctness and stopping iteration of GPGapE in finding an $(\epsilon, \delta)$-approximation to the top-$m$ data values. We further improve the computational efficiency, by calculating data values using small data subsets to reduce the computation cost of model training. We empirically demonstrate that GPGapE outperforms other baselines in top-$m$ data values identification, noisy data detection, and data subset selection on real-world datasets. We also demonstrate the efficiency of our GPGapE in data selection for large language model fine-tuning.

JBHI Journal 2025 Journal Article

FBCPM: A Filter Bank Connectome-Based Predictive Modeling Framework for EEG Signals

  • Linze Qian
  • Sujie Wang
  • Ioannis Kakkos
  • Xiaoyu Li
  • Xinyi Xu
  • Mengru Xu
  • George K. Matsopoulos
  • Yi Sun

The human brain connectome has long been recognized as a crucial component for various cognitive functions. While connectome-based predictive modeling (CPM) has been extensively explored for predicting behavior outcomes at the individual-level, its application to electroencephalogram (EEG) remains limited due to the inherent diversity and complexity of EEG frequency information. In the present work, we aim to address this issue by developing a filter bank CPM (FBCPM) framework that leverages narrowband EEG functional connectivity (FC) for individual prediction. Four independent datasets comprising 280 healthy subjects with 392 EEG recordings during the psychomotor vigilance test (PVT), were adopted here. Using the discovery dataset (i. e. , Dataset 1) with 137 recordings, the feasibility of FBCPM was evaluated via predicting mean reaction time (RT) measures within a 15-min PVT task. The results showed that FBCPM framework achieved notable prediction accuracy and outperformed four benchmark approaches. Subsequent comprehensive internal and external validation analyses further affirmed its robustness across various hyper-parameters and generalizability to another three independent datasets (i. e. , Dataset 2 to Dataset 4) with divergent recording or preprocessing settings. Moreover, the FBCPM framework exhibited satisfactory performance when generalized to time-on-task (TOT) effect measures (i. e. , $\mathit {\Delta RT}$ and $\mathit {TOT_{slope}}$ ). Further investigation of contributing features to mean RT prediction indicated the remarkable predictive ability of negative features, manifesting as a pattern of low-frequency (below 8 Hz) predominance and complex topological distributions. Overall, these findings indicated that FBCPM provided a significant methodological advance in EEG-based individual prediction approaches, moving a step forward towards practical application in cognitive neuroscience.

YNIMG Journal 2024 Journal Article

Age-dependent functional development pattern in neonatal brain: An fMRI-based brain entropy study

  • Zhiyong Zhao
  • Yifan Shuai
  • Yihan Wu
  • Xinyi Xu
  • Mingyang Li
  • Dan Wu

The relationship between brain entropy (BEN) and early brain development has been established through animal studies. However, it remains unclear whether the BEN can be used to identify age-dependent functional changes in human neonatal brains and the genetic underpinning of the new neuroimaging marker remains to be elucidated. In this study, we analyzed resting-state fMRI data from the Developing Human Connectome Project, including 280 infants who were scanned at 37.5-43.5 weeks postmenstrual age. The BEN maps were calculated for each subject, and a voxel-wise analysis was conducted using a general linear model to examine the effects of age, sex, and preterm birth on BEN. Additionally, we evaluated the correlation between regional BEN and gene expression levels. Our results demonstrated that the BEN in the sensorimotor-auditory and association cortices, along the 'S-A' axis, was significantly positively correlated with postnatal age (PNA), and negatively correlated with gestational age (GA), respectively. Meanwhile, the BEN in the right rolandic operculum correlated significantly with both GA and PNA. Preterm-born infants exhibited increased BEN values in widespread cortical areas, particularly in the visual-motor cortex, when compared to term-born infants. Moreover, we identified five BEN-related genes (DNAJC12, FIG4, STX12, CETN2, and IRF2BP2), which were involved in protein folding, synaptic vesicle transportation and cell division. These findings suggest that the fMRI-based BEN can serve as an indicator of age-dependent brain functional development in human neonates, which may be influenced by specific genes.

NeurIPS Conference 2024 Conference Paper

Data Distribution Valuation

  • Xinyi Xu
  • Shuaiqi Wang
  • Chuan-Sheng Foo
  • Bryan K. Low
  • Giulia Fanti

Data valuation is a class of techniques for quantitatively assessing the value of data for applications like pricing in data marketplaces. Existing data valuation methods define a value for a discrete dataset. However, in many use cases, users are interested in not only the value of the dataset, but that of the distribution from which the dataset was sampled. For example, consider a buyer trying to evaluate whether to purchase data from different vendors. The buyer may observe (and compare) only a small preview sample from each vendor, to decide which vendor's data distribution is most useful to the buyer and purchase. The core question is how should we compare the values of data distributions from their samples? Under a Huber characterization of the data heterogeneity across vendors, we propose a maximum mean discrepancy (MMD)-based valuation method which enables theoretically principled and actionable policies for comparing data distributions from samples. We empirically demonstrate that our method is sample-efficient and effective in identifying valuable data distributions against several existing baselines, on multiple real-world datasets (e. g. , network intrusion detection, credit card fraud detection) and downstream applications (classification, regression).

NeurIPS Conference 2024 Conference Paper

DETAIL: Task DEmonsTration Attribution for Interpretable In-context Learning

  • Zijian Zhou
  • Xiaoqiang Lin
  • Xinyi Xu
  • Alok Prakash
  • Daniela Rus
  • Bryan Kian Hsiang Low

In-context learning (ICL) allows transformer-based language models that are pre-trained on general text to quickly learn a specific task with a few "task demonstrations" without updating their parameters, significantly boosting their flexibility and generality. ICL possesses many distinct characteristics from conventional machine learning, thereby requiring new approaches to interpret this learning paradigm. Taking the viewpoint of recent works showing that transformers learn in context by formulating an internal optimizer, we propose an influence function-based attribution technique, DETAIL, that addresses the specific characteristics of ICL. We empirically verify the effectiveness of our approach for demonstration attribution while being computationally efficient. Leveraging the results, we then show how DETAIL can help improve model performance in real-world scenarios through demonstration reordering and curation. Finally, we experimentally prove the wide applicability of DETAIL by showing our attribution scores obtained on white-box models are transferable to black-box models in improving model performance.

ICML Conference 2024 Conference Paper

Distributionally Robust Data Valuation

  • Xiaoqiang Lin
  • Xinyi Xu
  • Zhaoxuan Wu
  • See-Kiong Ng
  • Bryan Kian Hsiang Low

Data valuation quantifies the contribution of each data point to the performance of a machine learning model. Existing works typically define the value of data by its improvement of the validation performance of the trained model. However, this approach can be impractical to apply in collaborative machine learning and data marketplace since it is difficult for the parties/buyers to agree on a common validation dataset or determine the exact validation distribution a priori. To address this, we propose a distributionally robust data valuation approach to perform data valuation without known/fixed validation distributions. Our approach defines the value of data by its improvement of the distributionally robust generalization error (DRGE), thus providing a worst-case performance guarantee without a known/fixed validation distribution. However, since computing DRGE directly is infeasible, we propose using model deviation as a proxy for the marginal improvement of DRGE (for kernel regression and neural networks) to compute data values. Furthermore, we identify a notion of uniqueness where low uniqueness characterizes low-value data. We empirically demonstrate that our approach outperforms existing data valuation approaches in data selection and data removal tasks on real-world datasets (e. g. , housing price prediction, diabetes hospitalization prediction).

ICML Conference 2023 Conference Paper

Collaborative Causal Inference with Fair Incentives

  • Rui Qiao 0006
  • Xinyi Xu
  • Bryan Kian Hsiang Low

Collaborative causal inference (CCI) aims to improve the estimation of the causal effect of treatment variables by utilizing data aggregated from multiple self-interested parties. Since their source data are valuable proprietary assets that can be costly or tedious to obtain, every party has to be incentivized to be willing to contribute to the collaboration, such as with a guaranteed fair and sufficiently valuable reward (than performing causal inference on its own). This paper presents a reward scheme designed using the unique statistical properties that are required by causal inference to guarantee certain desirable incentive criteria (e. g. , fairness, benefit) for the parties based on their contributions. To achieve this, we propose a data valuation function to value parties’ data for CCI based on the distributional closeness of its resulting treatment effect estimate to that utilizing the aggregated data from all parties. Then, we show how to value the parties’ rewards fairly based on a modified variant of the Shapley value arising from our proposed data valuation for CCI. Finally, the Shapley fair rewards to the parties are realized in the form of improved, stochastically perturbed treatment effect estimates. We empirically demonstrate the effectiveness of our reward scheme using simulated and real-world datasets.

YNIMG Journal 2023 Journal Article

Developmental pattern of individual morphometric similarity network in the human fetal brain

  • Ruoke Zhao
  • Cong Sun
  • Xinyi Xu
  • Zhiyong Zhao
  • Mingyang Li
  • Ruike Chen
  • Yao Shen
  • Yibin Pan

The development of the cerebral cortex during the fetal period is a complex yet well-coordinated process. MRI-based morphological brain network provides a powerful tool for describing this process at a network level. Due to the challenges of in-utero MRI acquisition and image processing, the fetal morphological brain network has not been established. In this study, utilizing high-resolution in-utero MRI data, we constructed an individual morphometric similarity network for each fetus based on multiple cortical features. The spatiotemporal development of morphological connections was described at the level of edge, node, and lobe, respectively. Based on graph theoretical method, the topology structure of fetal morphological network was characterized. Edge analysis demonstrated an increase of morphological dissimilarity between hemispheres with gestational age, especially for the parietal cortex. The limbic and parieto-occipital regions exhibited the most drastic changes of morphological connections at both the edge and node levels. Between- and within-lobe analysis illustrated that the limbic lobe became more similar to other lobes, while the parietal and occipital lobes became more dissimilar to other lobes. Graph theoretical analysis indicated that the small-world structure of the fetal morphological network appeared as early as 22 weeks and that the network topology exhibited an enhanced integration and reduced segregation during prenatal development. The findings obtained from the preterm-born neonates agreed well with those of the fetuses. In summary, this study fills a gap in prenatal morphological brain network research and provides a piece of important evidence for understanding the normal development of fetal brain connectome during the second-third trimester.

ICML Conference 2023 Conference Paper

Fair yet Asymptotically Equal Collaborative Learning

  • Xiaoqiang Lin
  • Xinyi Xu
  • See-Kiong Ng
  • Chuan-Sheng Foo
  • Bryan Kian Hsiang Low

In collaborative learning with streaming data, nodes (e. g. , organizations) jointly and continuously learn a machine learning (ML) model by sharing the latest model updates computed from their latest streaming data. For the more resourceful nodes to be willing to share their model updates, they need to be fairly incentivized. This paper explores an incentive design that guarantees fairness so that nodes receive rewards commensurate to their contributions. Our approach leverages an explore-then-exploit formulation to estimate the nodes’ contributions (i. e. , exploration) for realizing our theoretically guaranteed fair incentives (i. e. , exploitation). However, we observe a "rich get richer" phenomenon arising from the existing approaches to guarantee fairness and it discourages the participation of the less resourceful nodes. To remedy this, we additionally preserve asymptotic equality, i. e. , less resourceful nodes achieve equal performance eventually to the more resourceful/“rich” nodes. We empirically demonstrate in two settings with real-world streaming data: federated online incremental learning and federated reinforcement learning, that our proposed approach outperforms existing baselines in fairness and learning performance while remaining competitive in preserving equality.

NeurIPS Conference 2023 Conference Paper

Incentives in Private Collaborative Machine Learning

  • Rachael Sim
  • Yehong Zhang
  • Nghia Hoang
  • Xinyi Xu
  • Bryan Kian Hsiang Low
  • Patrick Jaillet

Collaborative machine learning involves training models on data from multiple parties but must incentivize their participation. Existing data valuation methods fairly value and reward each party based on shared data or model parameters but neglect the privacy risks involved. To address this, we introduce differential privacy (DP) as an incentive. Each party can select its required DP guarantee and perturb its sufficient statistic (SS) accordingly. The mediator values the perturbed SS by the Bayesian surprise it elicits about the model parameters. As our valuation function enforces a privacy-valuation trade-off, parties are deterred from selecting excessive DP guarantees that reduce the utility of the grand coalition's model. Finally, the mediator rewards each party with different posterior samples of the model parameters. Such rewards still satisfy existing incentives like fairness but additionally preserve DP and a high similarity to the grand coalition's posterior. We empirically demonstrate the effectiveness and practicality of our approach on synthetic and real-world datasets.

NeurIPS Conference 2023 Conference Paper

Model Shapley: Equitable Model Valuation with Black-box Access

  • Xinyi Xu
  • Thanh Lam
  • Chuan Sheng Foo
  • Bryan Kian Hsiang Low

Valuation methods of data and machine learning (ML) models are essential to the establishment of AI marketplaces. Importantly, certain practical considerations (e. g. , operational constraints, legal restrictions) favor the use of model valuation over data valuation. Also, existing marketplaces that involve trading of pre-trained ML models call for an equitable model valuation method to price them. In particular, we investigate the black-box access setting which allows querying a model (to observe predictions) without disclosing model-specific information (e. g. , architecture and parameters). By exploiting a Dirichlet abstraction of a model’s predictions, we propose a novel and equitable model valuation method called model Shapley. We also leverage a Lipschitz continuity of model Shapley to design a learning approach for predicting the model Shapley values (MSVs) of many vendors’ models (e. g. , 150) in a large-scale marketplace. We perform extensive empirical validation on the effectiveness of model Shapley using various real-world datasets and heterogeneous model types.

YNIMG Journal 2023 Journal Article

Multi-modal multi-resolution atlas of the human neonatal cerebral cortex based on microstructural similarity

  • Mingyang Li
  • Xinyi Xu
  • Zuozhen Cao
  • Ruike Chen
  • Ruoke Zhao
  • Zhiyong Zhao
  • Xixi Dang
  • Kenichi Oishi

The neonatal period is a critical window for the development of the human brain and may hold implications for the long-term development of cognition and disorders. Multi-modal connectome studies have revealed many important findings underlying the adult brain but related studies were rare in the early human brain. One potential challenge is the lack of an appropriate and unbiased parcellation that combines structural and functional information in this population. Using 348 multi-modal MRI datasets from the developing human connectome project, we found that the information fused from the structural, diffusion, and functional MRI was relatively stable across MRI features and showed high reproducibility at the group level. Therefore, we generated automated multi-resolution parcellations (300 - 500 parcels) based on the similarity across multi-modal features using a gradient-based parcellation algorithm. In addition, to acquire a parcellation with high interpretability, we provided a manually delineated parcellation (210 parcels), which was approximately symmetric, and the adjacent areas around each boundary were statistically different in terms of the integrated similarity metric and at least one kind of original features. Overall, the present study provided multi-resolution and neonate-specific parcellations of the cerebral cortex based on multi-modal MRI properties, which may facilitate future studies of the human connectome in the early development period.

AAAI Conference 2023 Conference Paper

Probably Approximate Shapley Fairness with Applications in Machine Learning

  • Zijian Zhou
  • Xinyi Xu
  • Rachael Hwee Ling Sim
  • Chuan Sheng Foo
  • Bryan Kian Hsiang Low

The Shapley value (SV) is adopted in various scenarios in machine learning (ML), including data valuation, agent valuation, and feature attribution, as it satisfies their fairness requirements. However, as exact SVs are infeasible to compute in practice, SV estimates are approximated instead. This approximation step raises an important question: do the SV estimates preserve the fairness guarantees of exact SVs? We observe that the fairness guarantees of exact SVs are too restrictive for SV estimates. Thus, we generalise Shapley fairness to probably approximate Shapley fairness and propose fidelity score, a metric to measure the variation of SV estimates, that determines how probable the fairness guarantees hold. Our last theoretical contribution is a novel greedy active estimation (GAE) algorithm that will maximise the lowest fidelity score and achieve a better fairness guarantee than the de facto Monte-Carlo estimation. We empirically verify GAE outperforms several existing methods in guaranteeing fairness while remaining competitive in estimation accuracy in various ML scenarios using real-world datasets.

IJCAI Conference 2022 Conference Paper

Data Valuation in Machine Learning: " Ingredients" , Strategies, and Open Challenges

  • Rachael Hwee Ling Sim
  • Xinyi Xu
  • Bryan Kian Hsiang Low

Data valuation in machine learning (ML) is an emerging research area that studies the worth of data in ML. Data valuation is used in collaborative ML to determine a fair compensation for every data owner and in interpretable ML to identify the most responsible, noisy, or misleading training examples. This paper presents a comprehensive technical survey that provides a new formal study of data valuation in ML through its “ingredients” and the corresponding properties, grounds the discussion of common desiderata satisfied by existing data valuation strategies on our proposed ingredients, and identifies open research challenges for designing new ingredients, data valuation strategies, and cost reduction techniques.

AAAI Conference 2022 Conference Paper

Incentivizing Collaboration in Machine Learning via Synthetic Data Rewards

  • Sebastian Shenghong Tay
  • Xinyi Xu
  • Chuan Sheng Foo
  • Bryan Kian Hsiang Low

This paper presents a novel collaborative generative modeling (CGM) framework that incentivizes collaboration among self-interested parties to contribute data to a pool for training a generative model (e. g. , GAN), from which synthetic data are drawn and distributed to the parties as rewards commensurate to their contributions. Distributing synthetic data as rewards (instead of trained models or money) offers taskand model-agnostic benefits for downstream learning tasks and is less likely to violate data privacy regulation. To realize the framework, we firstly propose a data valuation function using maximum mean discrepancy (MMD) that values data based on its quantity and quality in terms of its closeness to the true data distribution and provide theoretical results guiding the kernel choice in our MMD-based data valuation function. Then, we formulate the reward scheme as a linear optimization problem that when solved, guarantees certain incentives such as fairness in the CGM framework. We devise a weighted sampling algorithm for generating synthetic data to be distributed to each party as reward such that the value of its data and the synthetic data combined matches its assigned reward value by the reward scheme. We empirically show using simulated and real-world datasets that the parties’ synthetic data rewards are commensurate to their contributions.

ICML Conference 2022 Conference Paper

On the Convergence of the Shapley Value in Parametric Bayesian Learning Games

  • Lucas Agussurja
  • Xinyi Xu
  • Bryan Kian Hsiang Low

Measuring contributions is a classical problem in cooperative game theory where the Shapley value is the most well-known solution concept. In this paper, we establish the convergence property of the Shapley value in parametric Bayesian learning games where players perform a Bayesian inference using their combined data, and the posterior-prior KL divergence is used as the characteristic function. We show that for any two players, under some regularity conditions, their difference in Shapley value converges in probability to the difference in Shapley value of a limiting game whose characteristic function is proportional to the log-determinant of the joint Fisher information. As an application, we present an online collaborative learning framework that is asymptotically Shapley-fair. Our result enables this to be achieved without any costly computations of posterior-prior KL divergences. Only a consistent estimator of the Fisher information is needed. The effectiveness of our framework is demonstrated with experiments using real-world data.

NeurIPS Conference 2021 Conference Paper

Gradient Driven Rewards to Guarantee Fairness in Collaborative Machine Learning

  • Xinyi Xu
  • Lingjuan Lyu
  • Xingjun Ma
  • Chenglin Miao
  • Chuan Sheng Foo
  • Bryan Kian Hsiang Low

In collaborative machine learning(CML), multiple agents pool their resources(e. g. , data) together for a common learning task. In realistic CML settings where the agents are self-interested and not altruistic, they may be unwilling to share data or model information without adequate rewards. Furthermore, as the data/model information shared by the agents may differ in quality, designing rewards which are fair to them is important so that they would not feel exploited nor discouraged from sharing. In this paper, we adopt federated learning as the CML paradigm, propose a novel cosine gradient Shapley value(CGSV) to fairly evaluate the expected marginal contribution of each agent’s uploaded model parameter update/gradient without needing an auxiliary validation dataset, and based on the CGSV, design a novel training-time gradient reward mechanism with a fairness guarantee by sparsifying the aggregated parameter update/gradient downloaded from the server as reward to each agent such that its resulting quality is commensurate to that of the agent’s uploaded parameter update/gradient. We empirically demonstrate the effectiveness of our fair gradient reward mechanism on multiple benchmark datasets in terms of fairness, predictive performance, and time overhead.

NeurIPS Conference 2021 Conference Paper

Validation Free and Replication Robust Volume-based Data Valuation

  • Xinyi Xu
  • Zhaoxuan Wu
  • Chuan Sheng Foo
  • Bryan Kian Hsiang Low

Data valuation arises as a non-trivial challenge in real-world use cases such as collaborative machine learning, federated learning, trusted data sharing, data marketplaces. The value of data is often associated with the learning performance (e. g. , validation accuracy) of a model trained on the data, which introduces a close coupling between data valuation and validation. However, a validation set may notbe available in practice and it can be challenging for the data providers to reach an agreement on the choice of the validation set. Another practical issue is that of data replication: Given the value of some data points, a dishonest data provider may replicate these data points to exploit the valuation for a larger reward/payment. We observe that the diversity of the data points is an inherent property of a dataset that is independent of validation. We formalize diversity via the volume of the data matrix (i. e. , determinant of its left Gram), which allows us to establish a formal connection between the diversity of data and learning performance without requiring validation. Furthermore, we propose a robust volume measure with a theoretical guarantee on the replication robustness by following the intuition that copying the same data points does not increase the diversity of data. We perform extensive experiments to demonstrate its consistency in valuation and practical advantages over existing baselines and show that our method is model- and task-agnostic and can be flexibly adapted to handle various neural networks.

PRL Workshop 2020 Workshop Paper

Hierarchical Reinforcement Learning in StarCraft II with Human Expertise in Subgoals Selection

  • Xinyi Xu
  • Tiancheng Huang
  • Pengfei Wei
  • Akshay Narayan
  • Tze-Yun Leong

This work is inspired by recent advances in hierarchical reinforcement learning (HRL) (Barto and Mahadevan 2003; Hengst 2010), and improvements in learning efficiency from heuristic-based subgoal selection, experience replay (Lin 1993; Andrychowicz et al. 2017), and task-based curriculum learning (Bengio et al. 2009; Zaremba and Sutskever 2014). We propose a new method to integrate HRL, experience replay and effective subgoal selection through an implicit curriculum design based on human expertise to support sample-efficient learning and enhance interpretability of the agent’s behavior. Human expertise remains indispensable in many areas such as medicine (Buch, Ahmed, and Maruthappu 2018) and law (Cath 2018), where interpretability, explainability and transparency are crucial in the decision making process, for ethical and legal reasons. Our method simplifies the complex task sets for achieving the overall objectives by decomposing them into subgoals at different levels of abstraction. Incorporating relevant subjective knowledge also significantly reduces the computational resources spent in exploration for RL, especially in high speed, changing, and complex environments where the transition dynamics cannot be effectively learned and modelled in a short time. Experimental results in two StarCraft II (SC2) (Vinyals et al. 2017) minigames demonstrate that our method can achieve better sample efficiency than flat and end-to-end RL methods, and provides an effective method for explaining the agent’s performance.

IJCAI Conference 2019 Conference Paper

Zero-shot Metric Learning

  • Xinyi Xu
  • Huanhuan Cao
  • Yanhua Yang
  • Erkun Yang
  • Cheng Deng

In this work, we tackle the zero-shot metric learning problem and propose a novel method abbreviated as ZSML, with the purpose to learn a distance metric that measures the similarity of unseen categories (even unseen datasets). ZSML achieves strong transferability by capturing multi-nonlinear yet continuous relation among data. It is motivated by two facts: 1) relations can be essentially described from various perspectives; and 2) traditional binary supervision is insufficient to represent continuous visual similarity. Specifically, we first reformulate a collection of specific-shaped convolutional kernels to combine data pairs and generate multiple relation vectors. Furthermore, we design a new cross-update regression loss to discover continuous similarity. Extensive experiments including intra-dataset transfer and inter-dataset transfer on four benchmark datasets demonstrate that ZSML can achieve state-of-the-art performance.

IJCAI Conference 2018 Conference Paper

Deep View-Aware Metric Learning for Person Re-Identification

  • Pu Chen
  • Xinyi Xu
  • Cheng Deng

Person re-identification remains a challenging issue due to the dramatic changes in visual appearance caused by the variations in camera views, human pose, and background clutter. In this paper, we propose a deep view-aware metric learning (DVAML) model, where image pairs with similar and dissimilar views are projected into different feature subspaces, which can discover the intrinsic relevance between image pairs from different aspects. Additionally, we employ multiple metrics to jointly learn feature subspaces on which the relevance between image pairs are explicitly captured and thus greatly promoting the retrieval accuracy. Extensive experiment results on datasets CUHK01, CUHK03, and PRID2011 demonstrate the superiority of our method compared with state-of-the-art approaches.