Content
summary Summary

With T-FREE, a research team presents a new method that does not require a tokenizer and could significantly increase the efficiency of large language models.

Ad

Scientists from Aleph Alpha, the Technical University of Darmstadt, hessian.AI, and the German Research Center for Artificial Intelligence (DFKI) created T-FREE, which stands for "Tokenizer-Free Sparse Representations for Memory-Efficient Embeddings." Instead of using traditional tokenization, T-FREE directly embeds words through sparse activation patterns using character triples, which the team calls "trigrams." This approach allows for substantial compression of the embedding layer, which is responsible for converting text into numerical representations.

Image: Deiseroth et al.

In initial tests, the researchers achieved a parameter reduction of over 85 percent in these layers using T-FREE, without compromising performance in downstream tasks such as text classification or question-answering systems.

T-FREE facilitates transfer learning

One of the main advantages of T-FREE is its explicit modeling of morphological similarities between words. This means that similar word forms like "house," "houses," and "domestic" can be represented more efficiently in the model, as their similarities are directly incorporated into the coding. The researchers argue that the embeddings of such similar words should remain close to each other and can therefore be highly compressed. As a result, T-FREE can not only reduce the size of the embedding layers but also decrease the average encoding length of the text by 56 percent.

Ad
Ad

Additionally, T-FREE shows significant improvements in transfer learning between different languages. In an experiment with a 3-billion-parameter model trained first on English and then on German, T-FREE demonstrated much better adaptability than conventional tokenizer-based approaches.

However, the researchers acknowledge some limitations of their study. For example, the experiments have so far only been conducted with models of up to 3 billion parameters. Evaluations with larger models and training datasets are planned for the future.

More information about T-FREE is available on GitHub.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • Researchers from Aleph Alpha, TU Darmstadt, hessian.AI and DFKI have developed T-FREE, a new method for language modeling without a classical tokenizer. Instead, it uses direct embedding of words by sparse activation patterns over character triples.
  • In initial tests, T-FREE achieved a parameter reduction of over 85 percent in the embedding layers without compromising performance in tasks such as text classification or question-answer systems. In addition, the average coding length of the text was reduced by 56 percent.
  • T-FREE showed advantages in transfer learning between languages. In an experiment with a 3-billion-parameter model trained first on English and then on German, T-FREE proved to be significantly more adaptable than conventional tokenizer-based approaches.
Sources
Max is managing editor at THE DECODER. As a trained philosopher, he deals with consciousness, AI, and the question of whether machines can really think or just pretend to.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.