Portrait de Cheng Yang n'est pas disponible

Cheng Yang

Alumni

Publications

Latent Action Reparameterization for Efficient Agent Inference
Qingwen Zeng
Zerui Xu
Zijie Guo
Yu Sun
Siru Ouyang
Jiri Gesi
Fang Wu
Jiayi Zhang
Chenglin Wu
Xiangru Tang
Large language model (LLM) agents often rely on long sequences of low-level textual actions, resulting in large effective decision horizons … (voir plus)and high inference cost. While prior work has focused on improving inference efficiency through system-level optimizations or prompt engineering, we argue that a key bottleneck lies in the representation of the action space itself. We propose Latent Action Reparameterization (LAR), a framework that learns a compact latent action space in which each latent action corresponds to a multi-step semantic behavior. By reparameterizing agent actions into latent units, LAR enables decision making over a shorter effective horizon while preserving the expressiveness of the original action space. Unlike hand-crafted macros or hierarchical controllers, latent actions are learned from agent trajectories and integrated directly into the model, allowing both planning and execution to operate over abstract action representations. Across a range of LLM-based agent benchmarks, LAR significantly reduces the effective action horizon and improves inference efficiency under fixed compute budgets. As a consequence, our approach achieves substantial reductions in action tokens and corresponding wall-clock inference time, while maintaining or improving task success rates. These results suggest that action representation learning is a critical and underexplored factor in scaling efficient LLM agent inference, complementary to advances in model architecture and hardware.
Full-Scale Information Diffusion Prediction With Reinforced Recurrent Networks
Hao Wang
Chuan Shi
Maosong Sun
Ganqu Cui
Zhiyuan Liu
Information diffusion prediction is an important task, which studies how information items spread among users. With the success of deep lear… (voir plus)ning techniques, recurrent neural networks (RNNs) have shown their powerful capability in modeling information diffusion as sequential data. However, previous works focused on either microscopic diffusion prediction, which aims at guessing who will be the next influenced user at what time, or macroscopic diffusion prediction, which estimates the total numbers of influenced users during the diffusion process. To the best of our knowledge, few attempts have been made to suggest a unified model for both microscopic and macroscopic scales. In this article, we propose a novel full-scale diffusion prediction model based on reinforcement learning (RL). RL incorporates the macroscopic diffusion size information into the RNN-based microscopic diffusion model by addressing the nondifferentiable problem. We also employ an effective structural context extraction strategy to utilize the underlying social graph information. Experimental results show that our proposed model outperforms state-of-the-art baseline models on both microscopic and macroscopic diffusion predictions on three real-world datasets.