Arrow Research search

Author name cluster

Wei Yang Bryan Lim

Possible papers associated with this exact author name in Arrow. This page groups case-insensitive exact name matches and is not a full identity disambiguation profile.

10 papers
2 author rows

Possible papers

10

AAAI Conference 2026 Conference Paper

M-Loss: Quantifying Model Merging Compatibility with Limited Unlabeled Data

  • Tiantong Wang
  • Yiyang Duan
  • Haoyu Chen
  • Tiantong Wu
  • Wei Yang Bryan Lim

Training of large-scale models is both computationally intensive and often constrained by the availability of labeled data. Model merging offers a compelling alternative by directly integrating the weights of multiple source models without requiring additional data or extensive training. However, conventional model merging techniques, such as parameter averaging, often suffer from the unintended combination of non-generalizable features, especially when source models exhibit significant weight disparities. Comparatively, model ensembling generally provides more stable and superior performance that aggregates multiple models by averaging outputs. However, it incurs higher inference costs and increased storage requirements. While previous studies experimentally showed the similarities between model merging and ensembling, theoretical evidence and evaluation metrics remain lacking. To address this gap, we introduce Merging-ensembling loss (M-loss), a novel evaluation metric that quantifies the compatibility of merging source models using very limited unlabeled data. By measuring the discrepancy between parameter averaging and model ensembling at layer and node levels, M-loss facilitates more effective merging strategies. Specifically, M-loss serves both as a quantitative criterion of the theoretical feasibility of model merging, and a guide for parameter significance in model pruning. Our theoretical analysis and empirical evaluations demonstrate that incorporating M-loss into the merging process significantly improves the alignment between merged models and model ensembling, providing a scalable and efficient framework for accurate model consolidation.

AAAI Conference 2026 Conference Paper

Oblivionis: A Lightweight Learning and Unlearning Framework for Federated Large Language Models

  • Fuyao Zhang
  • Xinyu Yan
  • Tiantong Wu
  • Wenjie Li
  • Tianxiang Chen
  • Yang Cao
  • Ran Yan
  • Longtao Huang

Large Language Models (LLMs) increasingly leverage Federated Learning (FL) to utilize private, task-specific datasets for fine-tuning while preserving data privacy. However, while federated LLM frameworks effectively enable collaborative training without raw data sharing, they critically lack built-in mechanisms for regulatory compliance like GDPR’s right to be forgotten. Integrating private data heightens concerns over data quality and long-term governance, yet existing distributed training frameworks offer no principled way to selectively remove specific client contributions post-training. Due to distributed data silos, stringent privacy constraints, and the intricacies of interdependent model aggregation, federated LLM unlearning is significantly more complex than centralized LLM unlearning. To address this gap, we introduce Oblivionis, a lightweight learning and unlearning framework that enables clients to selectively remove specific private data during federated LLM training, enhancing trustworthiness and regulatory compliance. By unifying FL and unlearning as a dual optimization objective, we incorporate 6 FL and 5 unlearning algorithms for comprehensive evaluation and comparative analysis, establishing a robust pipeline for federated LLM unlearning. Extensive experiments demonstrate that Oblivionis outperforms local training, achieving a robust balance between forgetting efficacy and model utility, with cross-algorithm comparisons providing clear directions for future LLM development.

NeurIPS Conference 2025 Conference Paper

AegisGuard: RL-Guided Adapter Tuning for TEE-Based Efficient & Secure On-Device Inference

  • CHE WANG
  • Ziqi Zhang
  • Yinggui Wang
  • Tiantong Wang
  • Yurong Hao
  • Jianbo Gao
  • Tao Wei
  • Yang Cao

On-device large models (LMs) reduce cloud dependency but expose proprietary model weights to the end-user, making them vulnerable to white-box model stealing (MS) attacks. A common defense is TEE-Shielded DNN Partition (TSDP), which places all trainable LoRA adapters (fine tuned on private data) inside a trusted execution environment (TEE). However, this design suffers from excessive host-to-TEE communication latency. We propose AegisGuard, a fine tuning and deployment framework that selectively shields the MS sensitive adapters while offloading the rest to the GPU, balancing security and efficiency. AegisGuard integrates two key components: i) RL-based Sensitivity Measurement (RSM), which injects Gaussian noise during training and applies a lightweight reinforcement learning to rank adapters based on their impact on model stealing; and (ii) Shielded-Adapter Compression (SAC), which structurally prunes the selected adapters to reduce both parameter size and intermediate feature maps, further lowering TEE computation and data transfer costs. Extensive experiments demonstrate that AegisGuard achieves black-box level MS resilience (surrogate accuracy around 39%, matching fully shielded baselines), while reducing end-to-end inference latency by 2–3× and cutting TEE memory usage by 4× compared to state-of-the-art TSDP methods.

ICLR Conference 2025 Conference Paper

Enhancing Federated Domain Adaptation with Multi-Domain Prototype-Based Federated Fine-Tuning

  • Jingyuan Zhang
  • Yiyang Duan
  • Shuaicheng Niu
  • Yang Cao 0011
  • Wei Yang Bryan Lim

Federated Domain Adaptation (FDA) is a Federated Learning (FL) scenario where models are trained across multiple clients with unique data domains but a shared category space, without transmitting private data. The primary challenge in FDA is data heterogeneity, which causes significant divergences in gradient updates when using conventional averaging-based aggregation methods, reducing the efficacy of the global model. This further undermines both in-domain and out-of-domain performance (within the same federated system but outside the local client), which is critical in certain business applications. To address this, we propose a novel framework called \textbf{M}ulti-domain \textbf{P}rototype-based \textbf{F}ederated Fine-\textbf{T}uning (MPFT). MPFT fine-tunes a pre-trained model using multi-domain prototypes, i.e., several pretrained representations enriched with domain-specific information from category-specific local data. This enables supervised learning on the server to create a globally optimized adapter that is subsequently distributed to local clients, without the intrusion of data privacy. Empirical results show that MPFT significantly improves both in-domain and out-of-domain accuracy over conventional methods, enhancing knowledge preservation and adaptation in FDA. Notably, MPFT achieves convergence within a single communication round, greatly reducing computation and communication costs. To ensure privacy, MPFT applies differential privacy to protect the prototypes. Additionally, we develop a prototype-based feature space hijacking attack to evaluate robustness, confirming that raw data samples remain unrecoverable even after extensive training epochs. The complete implementation of MPFL is available at \url{https://anonymous.4open.science/r/DomainFL/}.

NeurIPS Conference 2025 Conference Paper

FedRAM: Federated Reweighting and Aggregation for Multi-Task Learning

  • Fan Wu
  • Xinyu Yan
  • Jiabei Liu
  • Wei Yang Bryan Lim

Federated Multi-Task Learning (FL-MTL) enables clients with heterogeneous data to collaboratively train models capable of handling multiple downstream tasks. However, FL-MTL faces key challenges, including statistical heterogeneity, task interference, and the need to balance local learning with global knowledge sharing. Traditional methods like FedAvg struggle in such settings due to the lack of explicit mechanisms to address these issues. In this paper, we propose FedRAM, a three-step framework that progressively updates two scalar hyperparameters: the task importance weight and the client aggregation coefficient. FedRAM introduces a reference-proxy-agent strategy, where the proxy model serves as an intermediate between the local reference model and the global agent model. This design reduces the need for repeated local training while preserving local performance. Extensive experiments on six real-world FL-MTL benchmarks show that FedRAM improves performance by at least 3$\%$ over the most baseline on both in-domain and out-of-domain tasks, while reducing computational cost by 15$\times$. These results make FedRAM a robust and practical solution for large-scale FL-MTL applications. The code is available at \url{https: //github. com/wwffvv/FedRAM}.

AAAI Conference 2022 System Paper

CrowdFL: A Marketplace for Crowdsourced Federated Learning

  • Daifei Feng
  • Cicilia Helena
  • Wei Yang Bryan Lim
  • Jer Shyuan Ng
  • Hongchao Jiang
  • Zehui Xiong
  • Jiawen Kang
  • Han Yu

Amid data privacy concerns, Federated Learning (FL) has emerged as a promising machine learning paradigm that enables privacy-preserving collaborative model training. However, there exists a need for a platform that matches data owners (supply) with model requesters (demand). In this paper, we present CrowdFL, a platform to facilitate the crowdsourcing of FL model training. It coordinates client selection, model training, and reputation management, which are essential steps for the FL crowdsourcing operations. By implementing model training on actual mobile devices, we demonstrate that the platform improves model performance and training efficiency. To the best of our knowledge, it is the first platform to support crowdsourcing-based FL on edge devices.

AAAI Conference 2022 System Paper

Dynamic Incentive Mechanism Design for COVID-19 Social Distancing

  • Xuan Rong Zane Ho
  • Wei Yang Bryan Lim
  • Hongchao Jiang
  • Jer Shyuan Ng
  • Han Yu
  • Zehui Xiong
  • Dusit Niyato
  • Chunyan Miao

As countries enter the endemic phase of COVID-19, people’s risk of exposure to the virus is greater than ever. There is a need to make more informed decisions in our daily lives on avoiding crowded places. Crowd monitoring systems typically require costly infrastructure. We propose a crowdsourced crowd monitoring platform which leverages user inputs to generate crowd counts and forecast location crowdedness. A key challenge for crowd-sourcing is a lack of incentive for users to contribute. We propose a Reinforcement Learning based dynamic incentive mechanism to optimally allocate rewards to encourage user participation.

AAAI Conference 2021 System Paper

AI-Empowered Decision Support for COVID-19 Social Distancing

  • Hongchao Jiang
  • Wei Yang Bryan Lim
  • Jer Shyuan Ng
  • Harold Ze Chie Teng
  • Han Yu
  • Zehui Xiong
  • Dusit Niyato
  • Chunyan Miao

The COVID-19 pandemic is one of the most severe challenges the world faces today. In order to contain the transmission of COVID-19, people around the world have been advised to practise social distancing. However, maintaining social distance is a challenging problem, as we often do not know beforehand how crowded the places we intend to visit are. In this paper, we demonstrate crowded. sg, an AIempowered platform that leverages on Unmanned Aerial Vehicles (UAVs), crowdsourced images, and computer vision techniques to provide social distancing decision support.

IJCAI Conference 2021 Conference Paper

Communication-efficient and Scalable Decentralized Federated Edge Learning

  • Austine Zong Han Yapp
  • Hong Soo Nicholas Koh
  • Yan Ting Lai
  • Jiawen Kang
  • Xuandi Li
  • Jer Shyuan Ng
  • Hongchao Jiang
  • Wei Yang Bryan Lim

Federated Edge Learning (FEL) is a distributed Machine Learning (ML) framework for collaborative training on edge devices. FEL improves data privacy over traditional centralized ML model training by keeping data on the devices and only sending local model updates to a central coordinator for aggregation. However, challenges still remain in existing FEL architectures where there is high communication overhead between edge devices and the coordinator. In this paper, we present a working prototype of blockchain-empowered and communication-efficient FEL framework, which enhances the security and scalability towards large-scale implementation of FEL.

IJCAI Conference 2021 Conference Paper

Predictive Analytics for COVID-19 Social Distancing

  • Harold Ze Chie Teng
  • Hongchao Jiang
  • Xuan Rong Zane Ho
  • Wei Yang Bryan Lim
  • Jer Shyuan Ng
  • Han Yu
  • Zehui Xiong
  • Dusit Niyato

The COVID-19 pandemic has disrupted the lives of millions across the globe. In Singapore, promoting safe distancing by managing crowds in public areas have been the cornerstone of containing the community spread of the virus. One of the most important solutions to maintain social distancing is to monitor the crowdedness of indoor and outdoor points of interest. Using Nanyang Technological University (NTU) as a testbed, we develop and deploy a platform that provides live and predicted crowd counts for key locations on campus to help users plan their trips in an informed manner, so as to mitigate the risk of community transmission.