Content
summary Summary

Yale University researchers have found that AI models learn best when trained on data that hits a specific complexity level – not too simple, not too chaotic.

Ad

The research team trained various language models using elementary cellular automata (ECAs), simple systems in which the future state of each cell depends on itself and its two neighbors. While these systems use basic rules, they can generate anything from simple to highly complex patterns. The researchers then studied the performance of these LLMs in reasoning tasks and in predicting chess moves.

Schematic diagram: Framework for studying complexity and intelligence using cellular automata, LLM training, and downstream tasks.
Image: Zhang, Patel

The study found that AI models trained on more complex ECA rules performed better in downstream tasks such as reasoning and predicting chess moves. Models trained on Class IV ECAs, according to Wolfram's classification, performed particularly well. The rules of these ECAs produce patterns that are neither completely ordered nor completely chaotic, but exhibit a kind of structured complexity.

The "Edge of Chaos"

"Surprisingly, we find that models can learn complex solutions when trained on simple rules. Our results point to an optimal complexity level, or 'edge of chaos', conducive to intelligence, where the system is structured yet challenging to predict," the authors say.

Ad
Ad

Models exposed to very simple patterns tended to learn trivial solutions, while those trained on more complex patterns developed more sophisticated abilities, even when simpler approaches were available. The researchers suspect that this complexity in the learned representations is a key factor enabling the models to transfer their knowledge to other tasks.

The findings could shed light on why large language models like GPT-3 and GPT-4 are so effective. According to the researchers, the sheer volume and diversity of training data used in these models might create benefits similar to those seen with complex ECA patterns in their study.

The team notes that more research is needed to verify this connection. They're planning to test their theory by expanding their experiments to include larger models and more complex systems.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • A study from Yale University suggests that the complexity of training data is critical to the development of intelligence in AI systems.
  • The researchers trained large language models on data from elementary cellular automata (ECAs) of varying complexity and tested their performance on reasoning tasks.
  • Models trained on the behavior of more complex ECA rules performed better on subsequent tasks. The results suggest an optimal level of complexity, or "edge of chaos," that fosters intelligence.
Sources
Max is managing editor at THE DECODER. As a trained philosopher, he deals with consciousness, AI, and the question of whether machines can really think or just pretend to.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.