AI-powered
podcast player
Listen to all your favourite podcasts with AI-powered features
Scaling Language Models and Reasoning
Scaling up language models like GPT-2 is unlikely to result in full-fledged reasoning abilities, as the type signature for thinking involves spending variable amounts of compute to arrive at better answers, a process not encoded in GPT-2. Small tweaks to the language model's process, such as generating whole sequences of thoughts and keeping only the final bit, may be necessary. Additionally, reasoning seems linked to out-of-distribution generalization, enabling refinement of mental models for unexperienced scenarios.