Arrow Research search

Author name cluster

Yitian Chen

Possible papers associated with this exact author name in Arrow. This page groups case-insensitive exact name matches and is not a full identity disambiguation profile.

2 papers
2 author rows

Possible papers

2

NeurIPS Conference 2025 Conference Paper

Solver-Informed RL: Grounding Large Language Models for Authentic Optimization Modeling

  • Yitian Chen
  • Jingfan Xia
  • Siyu Shao
  • Dongdong Ge
  • Yinyu Ye

Optimization modeling is fundamental to decision-making in fields such as supply chain management, logistics, and financial engineering, but its complexity presents a major barrier to adoption. Automating model creation from natural language is key to improving efficiency and access. However, while Large Language Models (LLMs) are a promising tool for this, they often produce flawed or infeasible results due to errors and hallucinations. To address this issue, we propose Solver-Informed Reinforcement Learning (SIRL), a framework that uses Reinforcement Learning with Verifiable Reward to improve LLMs’ ability to generate accurate and executable optimization models. Specifically, SIRL automatically assesses the executable code and the instance-level mathematical model represented by the associated. lp files. This process yields precise feedback on syntactic validity, feasibility, and solution quality, which serves as a direct reward signal to guide the reinforcement learning process. Furthermore, this verification mechanism also supports our instance-enhanced self-consistency method for creating high-quality training data. Extensive experiments on diverse public benchmarks demonstrate that models trained with our SIRL framework achieve state-of-the-art performance, substantially outperforming existing methods in generating accurate and executable optimization models. Specifically, our SIRL-32B model surpasses DeepSeek-V3 and OpenAI-o3 on the majority of these benchmarks. Our code is publicly available at https: //github. com/Cardinal-Operations/SIRL.

ECAI Conference 2024 Conference Paper

Matching Gains with Pays: Effective and Fair Learning in Multi-Agent Public Goods Dilemmas

  • Yitian Chen
  • Xuan Liu 0001
  • Shigeng Zhang
  • Xinning Chen
  • Song Guo 0001

The training of multi-agent reinforcement learning (MARL) tasks with the public goods dilemma (PGD) is difficult because the selfish actions of individual agents for high personal rewards may reduce the collective utility of the whole group. Existing solutions to this problem, e. g. , reward gifting or intrinsic rewards, although inducing cooperation among agents in small groups, cannot guarantee fairness among agents’ policies and fail to achieve optimal group utility in large-scale systems. In this paper, we propose F4PGD, an effective method to train large-scale MARL tasks with PGD in a decentralized manner, which is inspired by Adam’s equity theory that the match between a person’s payoff and his contribution is the key incentive for people to contribute to the common good. In F4PGD, a mechanism is designed to match an agent’s reward with its contribution, which suppresses agents from taking a free ride and meanwhile encourages well-learned agents to contribute to public goods. Experimental results show that F4PGD effectively learns optimal policies for the whole group and guarantees fairness among agents in several typical MARL tasks with PGD.