YuyaoGe's Website
YuyaoGe's Website
About
Highlight Publications
Other Publications
Projects
Posts
Light
Dark
Automatic
Home
Tags
Prompt
Prompt
Gated Differentiable Working Memory for Long-Context Language Modeling
Abstract: Long contexts challenge transformers: attention scores dilute across thousands of tokens, critical information is often lost in the middle, and models struggle to adapt to novel patterns at inference time.
Lingrui Mei
,
Shenghua Liu
,
Yiwei Wang
,
Yuyao Ge 葛钰峣
,
Baolong Bi
,
Jiayu Yao
,
Jun Wan
,
Ziling Yin
,
Jiafeng Guo
,
Xueqi Cheng
Jan 19, 2026
PDF
Cite
DOI
arXiv
Reward and Guidance through Rubrics: Promoting Exploration to Improve Multi-Domain Reasoning
Abstract: Recent advances in reinforcement learning (RL) have significantly improved the complex reasoning capabilities of large language models (LLMs). Despite these successes, existing methods mainly focus on single-domain RL (e.g., mathematics) with verifiable rewards (RLVR), and their reliance on purely online RL frameworks restricts the exploration space, thereby limiting reasoning performance.
Baolong Bi
,
Shenghua Liu
,
Yiwei Wang
,
Siqian Tong
,
Lingrui Mei
,
Yuyao Ge 葛钰峣
,
Yilong Xu
,
Jiafeng Guo
,
Xueqi Cheng
Nov 15, 2025
PDF
Cite
DOI
arXiv
Are All Prompt Components Value-Neutral? Understanding the Heterogeneous Adversarial Robustness of Dissected Prompt in Large Language Models
Abstract: Prompt-based adversarial attacks have become an effective means to assess the robustness of large language models (LLMs). However, existing approaches often treat prompts as monolithic text, overlooking their structural heterogeneity-different prompt components contribute unequally to adversarial robustness.
Yujia Zheng
,
Tianhao Li
,
Haotian Huang
,
Tianyu Zeng
,
Jingyu Lu
,
Chuangxin Chu
,
Yuekai Huang
,
Ziyou Jiang
,
Qian Xiong
,
Yuyao Ge 葛钰峣
,
Mingyang Li
Aug 3, 2025
PDF
Cite
DOI
arXiv
PIS:Linking Importance Sampling and Attention Mechanisms for Efficient Prompt Compression
Abstract: Large language models (LLMs) have achieved remarkable progress, demonstrating unprecedented capabilities across various natural language processing tasks. However, the high costs associated with such exceptional performance limit the widespread adoption of LLMs, highlighting the need for prompt compression.
Lizhe Chen
,
Binjia Zhou
,
Yuyao Ge 葛钰峣
,
Jiayi Chen
,
Shiguang Ni
Apr 23, 2025
PDF
Cite
DOI
arXiv
Cite
×