DE Alba
Subscribe
Reasoning
Language Models
Self-Taught
Quiet-STaR: Self-Taught Reasoning in Language Models

Quiet-STaR introduces a novel approach for language models to teach themselves reasoning by inferring underlying thought processes behind texts. Discover more in the full text.

  • The model generates rationales at each token to justify future text segments, enhancing model prediction accuracy.
  • A new sampling algorithm and learnable tokens are introduced to track the beginning and ending of thoughts.
  • Generated rationales significantly aid the model in addressing challenging tokens and incrementally improve performance on difficult questions.
  • Quiet-STaR’s zero-shot learning shows gains on benchmarks such as GSM8K and CommonsenseQA without task-specific fine-tuning.

Quiet-STaR’s self-teaching capabilities mark an advancement in teaching LMs to reason, suggesting a future where AI could autonomously learn and apply rational thought processes.

Personalized AI news from scientific papers.