Discussion about this post

User's avatar
Sahar Mor's avatar

The paper also includes some insights on how to prompt reasoning models:

(1) Zero-shot outperforms few-shot - Their extensive testing revealed that few-shot prompting consistently degrades model performance, contrary to traditional LLM best practices.

(2) Direct problem description wins - The model performs best when users simply state the problem and specify the output format, avoiding complex prompting patterns.

(3) Language consistency matters - Using the same language throughout the prompt is crucial, as the model can mix languages in reasoning chains when prompts contain multiple languages.

Expand full comment
Binit Mohanty's avatar

Just wanted to express my appreciation for this and all your previous posts. I have always found value in them and look forward to the next one.

Expand full comment
33 more comments...

No posts