Menu
in

Enhancing language models’ thinking ability with Quiet-STaR technology #AIimprovement

Researchers from Stanford University and Notbad AI, Inc. have developed Quiet-STaR, an extension of the Self-Taught Reasoner (STaR) model. This innovative AI model is trained to think before responding to prompts, similar to how humans consider their responses before speaking. By training on a wide corpus of internet data, Quiet-STaR learns to generate rationales at each token to explain future text and improve predictions. The model enhances zero-shot direct reasoning abilities, with improvements seen in question-answering challenges and math word problems.

Unlike previous AI reasoning methods that were more focused and less generalized, Quiet-STaR allows language models to reason generally from text, rather than being limited to specific datasets or tasks. The model generates inner thoughts in parallel to explain future text before responding to a prompt, producing a mixture of predictions with and without rationales. The researchers emphasize that Quiet-STaR’s training on diverse web text enables more robust and adaptable language models, bridging the gap between model and human reasoning capabilities.

The development of Quiet-STaR has significant potential for various applications, including enhancing reasoning abilities in the security workforce. An upcoming event in Atlanta, in partnership with Microsoft, will explore how generative AI is transforming security teams. Overall, Quiet-STaR represents a significant advancement in AI reasoning, with the potential to revolutionize language models and problem-solving capabilities. As research continues to refine and build upon these insights, the gap between language models and human-like reasoning capabilities will continue to narrow, unlocking new possibilities in various industries.

Source link

Source link: https://www.globalvillagespace.com/tech/how-quiet-star-enhances-language-models-ability-to-think-before-speaking/

Leave a Reply

Exit mobile version