Prompt Repetition Improves Non-Reasoning LLMs
<https://arxiv.org/pdf/2512.14982>
When not using reasoning, repeating the input prompt improves performance for popular models (Gemini, GPT, Claude, and Deepseek) without increasing the number of generated tokens or latency.
- Prompt Repetition LLMs are often trained as causal language models, i.e. past tokens cannot attend to future tokens. Therefore, the order of the tokens in a user’s query can affect prediction performance. For example, a query of the form “ ” often performs differently from a query of the form “ ” (see options-first vs. question-first in Figure 1). We propose to repeat the prompt, i.e. transform the input from “ ” to “ ”. This enables each prompt token to attend to every other prompt token, addressing the above. When not using reasoning, prompt repetition improves the performance of LLMs (Figure 1) without increasing the lengths of the generated outputs or latency
贡献者
longsizhuo贡献 1 次 · 最近 2026/05/06
TinyAlmond贡献 1 次 · 最近 2026/04/09
Mira190贡献 1 次 · 最近 2026/04/01
CeitherNSW贡献 1 次 · 最近 2026/03/05
最近更新
Involution Hell© 2026 byCommunityunderCC BY-NC-SA 4.0
LeWorldModel
Stable End-to-End Joint-Embedding Predictive Architecture from Pixels
Free AI Productivity Tools with Your Student Email
A series on AI productivity tools available for free with a student email, featuring in-depth usage tips for Perplexity Comet, Cursor, and V0.dev — covering AI-powered search, code generation, and frontend development. Ideal for current students, job-seeking developers, and CS/AI learners who want to maximize free resources and boost their productivity.