Math needs thinking time, everyday knowledge needs memory, and a new Transformer architecture aims to deliver both
A German research team lets Transformer models decide for themselves how many times they think about a problem. Combined with additional memory, the approach outperforms larger models on math problems.
Qualcomm shrinks AI reasoning chains by 2.4x to fit thinking models on smartphones
Qualcomm AI Research has developed a modular system that brings reasoning-capable language models to smartphones by compressing the models’ verbose thought processes by a factor of 2.4.