Arrow Research search

Author name cluster

Li Ding

Possible papers associated with this exact author name in Arrow. This page groups case-insensitive exact name matches and is not a full identity disambiguation profile.

17 papers
2 author rows

Possible papers

17

AAAI Conference 2026 Conference Paper

Model-Agnostic Sentiment Distribution Stability Analysis for Robust LLM-Generated Texts Detection

  • Siyuan Li
  • Xi Lin
  • Guangyan Li
  • Zehao Liu
  • Aodu Wulianghai
  • Li Ding
  • Jun Wu
  • Jianhua Li

The rapid advancement of large language models (LLMs) has resulted in increasingly sophisticated AI-generated content, posing significant challenges in distinguishing LLM-generated text from human-written language. Existing detection methods, primarily based on lexical heuristics or fine-tuned classifiers, often suffer from limited generalizability and are vulnerable to paraphrasing, adversarial perturbations, and cross-domain shifts. In this work, we propose SentiDetect, a model-agnostic framework for detecting LLM-generated text by analyzing the divergence in sentiment distribution stability. Our method is motivated by the empirical observation that LLM outputs tend to exhibit emotionally consistent patterns, whereas human-written texts display greater emotional variability. To capture this phenomenon, we define two complementary metrics: sentiment distribution consistency and sentiment distribution preservation, which quantify stability under sentiment-altering and semantic-preserving transformations. We evaluate SentiDetect on five diverse domains and a range of advanced LLMs, including Gemini-1.5-Pro, Claude-3, GPT-4-0613, and LLaMa-3.3. Experimental results demonstrate its superiority over state-of-the-art baselines, with over 16% and 11% F1 score improvements on Gemini-1.5-Pro and GPT-4-0613, respectively. Moreover, SentiDetect also shows greater robustness to paraphrasing, adversarial attacks, and text length variations, outperforming existing detectors in challenging scenarios.

ICML Conference 2025 Conference Paper

LBI-FL: Low-Bit Integerized Federated Learning with Temporally Dynamic Bit-Width Allocation

  • Li Ding
  • Hao Zhang
  • Wenrui Dai
  • Chenglin Li
  • Weijia Lu
  • Zhifei Yang 0005
  • Xiaodong Zhang
  • Xiaofeng Ma

Federated learning (FL) is greatly challenged by the communication bottleneck and computation limitation on clients. Existing methods based on quantization for FL cannot simultaneously reduce the uplink and downlink communication cost and mitigate the computation burden on clients. To address this problem, in this paper, we propose the first low-bit integerized federated learning (LBI-FL) framework that quantizes the weights, activations, and gradients to lower than INT8 precision to evidently reduce the communication and computational costs. Specifically, we achieve dynamical temporal bit-width allocation for weights, activations, and gradients along the training trajectory via reinforcement learning. An agent is trained to determine bit-width allocation by comprehensively considering the states like current bit-width, training stage, and quantization loss as the state. The agent efficiently trained on small-scale datasets can be well generalized to train varying network architectures on non-independent and identically distributed datasets. Furthermore, we demonstrated in theory that federated learning with gradient quantization achieves an equivalent convergence rate to FedAvg. The proposed LBI-FL can reduce the communication costs by 8 times compared to full-precision FL. Extensive experiments show that the proposed LBI-FL achieves a reduction of more than 50% BitOPs per client on average for FL with less than 2% accuracy loss compared to low-bit training with INT8 precision.

RLC Conference 2025 Conference Paper

Pareto Optimal Learning from Preferences with Hidden Context

  • Ryan Bahlous-Boldi
  • Li Ding
  • Lee Spector
  • Scott Niekum

Ensuring AI models align with human values is essential for their safety and functionality. Reinforcement learning from human feedback (RLHF) leverages human preferences to achieve this alignment. However, when preferences are sourced from diverse populations, point estimates of reward can result in suboptimal performance or be unfair to specific groups. We propose Pareto Optimal Preference Learning (POPL), which enables pluralistic alignment by framing discrepant group preferences as objectives with potential trade-offs, aiming for policies that are Pareto-optimal on the preference dataset. POPL utilizes lexicase selection, an iterative process that selects diverse and Pareto-optimal solutions. Our theoretical and empirical evaluations demonstrate that POPL surpasses baseline methods in learning sets of reward functions and policies, effectively catering to distinct groups without access to group numbers or membership labels. We verify the performance of POPL on a stateless preference learning setting, a Minigrid RL domain, Metaworld robotics benchmarks, as well as large language model (LLM) fine-tuning. We illustrate that POPL can also serve as a foundation for techniques optimizing specific notions of group fairness, ensuring safe and equitable AI model alignment.

RLJ Journal 2025 Journal Article

Pareto Optimal Learning from Preferences with Hidden Context

  • Ryan Bahlous-Boldi
  • Li Ding
  • Lee Spector
  • Scott Niekum

Ensuring AI models align with human values is essential for their safety and functionality. Reinforcement learning from human feedback (RLHF) leverages human preferences to achieve this alignment. However, when preferences are sourced from diverse populations, point estimates of reward can result in suboptimal performance or be unfair to specific groups. We propose Pareto Optimal Preference Learning (POPL), which enables pluralistic alignment by framing discrepant group preferences as objectives with potential trade-offs, aiming for policies that are Pareto-optimal on the preference dataset. POPL utilizes lexicase selection, an iterative process that selects diverse and Pareto-optimal solutions. Our theoretical and empirical evaluations demonstrate that POPL surpasses baseline methods in learning sets of reward functions and policies, effectively catering to distinct groups without access to group numbers or membership labels. We verify the performance of POPL on a stateless preference learning setting, a Minigrid RL domain, Metaworld robotics benchmarks, as well as large language model (LLM) fine-tuning. We illustrate that POPL can also serve as a foundation for techniques optimizing specific notions of group fairness, ensuring safe and equitable AI model alignment.

NeurIPS Conference 2025 Conference Paper

QiMeng-MuPa: Mutual-Supervised Learning for Sequential-to-Parallel Code Translation

  • Changxin Ke
  • Rui Zhang
  • Shuo Wang
  • Li Ding
  • Guangli Li
  • Yuanbo Wen
  • Shuoming Zhang
  • Ruiyuan Xu

The rise of GPU-based high-performance computing (HPC) has driven the widespread adoption of parallel programming models such as CUDA. Yet, the inherent complexity of parallel programming creates a demand for the automated sequential-to-parallel approaches. However, data scarcity poses a significant challenge for machine learning-based sequential-to-parallel code translation. Although recent back-translation methods show promise, they still fail to ensure functional equivalence in the translated code. In this paper, we propose \textbf{QiMeng-MuPa}, a novel \textbf{Mu}tual-Supervised Learning framework for Sequential-to-\textbf{Pa}rallel code translation, to address the functional equivalence issue. QiMeng-MuPa consists of two models, a Translator and a Tester. Through an iterative loop consisting of Co-verify and Co-evolve steps, the Translator and the Tester mutually generate data for each other and improve collectively. The Tester generates unit tests to verify and filter functionally equivalent translated code, thereby evolving the Translator, while the Translator generates translated code as augmented input to evolve the Tester. Experimental results demonstrate that QiMeng-MuPa significantly enhances the performance of the base models: when applied to Qwen2. 5-Coder, it not only improves Pass@1 by up to 28. 91\% and boosts Tester performance by 68. 90\%, but also outperforms the previous state-of-the-art method CodeRosetta by 1. 56 and 6. 92 in BLEU and CodeBLEU scores, while achieving performance comparable to DeepSeek-R1 and GPT-4. 1. Our code is available at \url{https: //github. com/kcxain/mupa}.

ICML Conference 2024 Conference Paper

AMPA: Adaptive Mixed Precision Allocation for Low-Bit Integer Training

  • Li Ding
  • Wen Fei
  • Yuyang Huang
  • Shuangrui Ding
  • Wenrui Dai
  • Chenglin Li
  • Junni Zou
  • Hongkai Xiong

Low-bit integer training emerges as a promising approach to mitigate the heavy burden during network training by quantizing the weights, activations, and gradients. However, existing methods cannot well achieve mixed-precision quantization for low-bit training and are commonly limited to INT8 precision. In this paper, we propose a novel low-bit integer training framework that, for the first time, achieves adaptive mixed-precision allocation (AMPA) for weights, activations, and gradients, and pushes the boundaries to a precision level below INT8. We develop a novel magnitude-based sensitivity measurement with regard to the quantization losses of weight, activation, and gradient quantization and the average gradient magnitudes, which is demonstrated as an upper bound of quantization influence in theory. We further design a layer-wise precision update strategy under observations on the quantization losses and their effects on model performance in low-bit training. Extensive experiments on different backbones and datasets show that, compared to INT8 quantization, the proposed method can achieve more than 38% BitOPs reduction with a tolerable loss below 2% in image classification, image segmentation, and language modeling.

AAAI Conference 2021 Conference Paper

Towards Universal Physical Attacks on Single Object Tracking

  • Li Ding
  • Yongwei Wang
  • Kaiwen Yuan
  • Minyang Jiang
  • Ping Wang
  • Hua Huang
  • Z. Jane Wang

Recent studies show that small perturbations in video frames could misguide single object trackers. However, such attacks have been mainly designed for digital-domain videos (i. e. , perturbation on full images), which makes them practically infeasible to evaluate the adversarial vulnerability of trackers in real-world scenarios. Here we made the first step towards physically feasible adversarial attacks against visual tracking in real scenes with a universal patch to camouflage single object trackers. Fundamentally different from physical object detection, the essence of single object tracking lies in the feature matching between the search image and templates, and we therefore specially design the maximum textural discrepancy (MTD), a resolution-invariant and target location-independent feature de-matching loss. The MTD distills global textural information of the template and search images at hierarchical feature scales prior to performing feature attacks. Moreover, we evaluate two shape attacks, the regression dilation and shrinking, to generate stronger and more controllable attacks. Further, we employ a set of transformations to simulate diverse visual tracking scenes in the wild. Experimental results show the effectiveness of the physically feasible attacks on SiamMask and SiamRPN++ visual trackers both in digital and physical scenes.

JBHI Journal 2020 Journal Article

Measuring and Localizing Individual Bites Using a Sensor Augmented Plate During Unrestricted Eating for the Aging Population

  • Gert Mertes
  • Li Ding
  • Wei Chen
  • Hans Hallez
  • Jie Jia
  • Bart Vanrumste

Food intake monitoring can play an important role in the prevention of malnutrition in the aging population, but traditional tools may not be adequate for use in this target group. These tools typically involve the use of questionnaires or food diaries that require manual data entry. Due to their time-consuming nature, they are often incomplete, contain mistakes, or not used at all. An alternative to self-reporting tools, in the form of a plate system that automatically measures the consumed food during the meal, is presented in this paper. Furthermore, the system can estimate the location where each bite was taken on the plate. The system is compatible with an off-the-shelf plate that is mounted on top of a base station. Weight sensors are integrated in the base, allowing for easy removal and cleaning of the plate. Localization of bites is done by looking at the movement of the center of mass during eating. When used with a compartmentalized plate, the amount of consumed food per compartment can be measured. With prior knowledge of the type of food in each compartment, this can give an indication of calories and nutritional intake. We present a bite detection algorithm using a random forest decision tree classifier. Data from 24 aging adults (ages 52-95) eating a single meal with chopsticks was used to train and evaluate the model. Out of a total of 836 true annotated bites, the algorithm detected 602 with a precision and recall of 0. 78 and 0. 76, respectively. By summing the weights of detected bites from each compartment, the algorithm was able to estimate the amount of food taken per compartment with an average error of $(8 \pm 8)$% of the portion size.

TIST Journal 2012 Journal Article

An Ensemble Architecture for Learning Complex Problem-Solving Techniques from Demonstration

  • Xiaoqin Shelley Zhang
  • Bhavesh Shrestha
  • Sungwook Yoon
  • Subbarao Kambhampati
  • Phillip DiBona
  • Jinhong K. Guo
  • Daniel McFarlane
  • Martin O. Hofmann

We present a novel ensemble architecture for learning problem-solving techniques from a very small number of expert solutions and demonstrate its effectiveness in a complex real-world domain. The key feature of our “Generalized Integrated Learning Architecture” (GILA) is a set of heterogeneous independent learning and reasoning (ILR) components, coordinated by a central meta-reasoning executive (MRE). The ILRs are weakly coupled in the sense that all coordination during learning and performance happens through the MRE. Each ILR learns independently from a small number of expert demonstrations of a complex task. During performance, each ILR proposes partial solutions to subproblems posed by the MRE, which are then selected from and pieced together by the MRE to produce a complete solution. The heterogeneity of the learner-reasoners allows both learning and problem solving to be more effective because their abilities and biases are complementary and synergistic. We describe the application of this novel learning and problem solving architecture to the domain of airspace management, where multiple requests for the use of airspaces need to be deconflicted, reconciled, and managed automatically. Formal evaluations show that our system performs as well as or better than humans after learning from the same training data. Furthermore, GILA outperforms any individual ILR run in isolation, thus demonstrating the power of the ensemble architecture for learning and problem solving.

IS Journal 2012 Journal Article

Linked Open Government Data [Guest editors' introduction]

  • Li Ding
  • Vassilios Peristeras
  • Michael Hausenblas

Government data covers authoritative and valuable information about our society. Public access to government data, however, remains challenging largely due to the heterogeneity and complexity of the public information ecosystem which results in high costs for locating, decoding, inter-linking and reusing existing government data. Recently, linked data–based solutions have been adopted by the leading practitioners (such as Data. gov in the US and Data. gov. uk in the UK) to offer an open and incremental ecosystem that interconnects providers, consumers, and contributors of open government data. This article first reports a community consensus on the architecture of the linked open government data ecosystem, then reviews the key technologies reported by works included in this special issue, and finally concludes with three grand challenges towards opening, linking, and reusing government data.

IS Journal 2005 Journal Article

Social Networks Applied

  • S. Staab
  • P. Domingos
  • P. Mika
  • J. Golbeck
  • Li Ding
  • T. Finin
  • A. Joshi
  • A. Nowak

Social networks have interesting properties. They influence our lives enormously without us being aware of the implications they raise. The authors investigate the following areas concerning social networks: how to exploit our unprecedented wealth of data and how we can mine social networks for purposes such as marketing campaigns; social networks as a particular form of influence, i. e. ., the way that people agree on terminology and this phenomenon's implications for the way we build ontologies and the Semantic Web; social networks as something we can discover from data; the use of social network information to offer a wealth of new applications such as better recommendations for restaurants, trustworthy email senders, or (maybe) blind dates; investigation of the richness and difficulty of harvesting FOAF (friend-of-a-friend) information; and by looking at how information processing is bound to social context, the resulting ways that network topology's definition determines its outcomes.