Content
summary Summary

OpenAI is expanding its custom AI training offerings with a new method called Reinforcement Fine-Tuning (RFT). The technique aims to create specialized o1 models that can perform complex technical tasks with minimal training examples.

Ad

The new approach works differently from traditional supervised fine-tuning. Instead of just learning to copy the style and tone of training data, models can develop new ways of "thinking" through problems, according to OpenAI. When given a problem, the model gets time to work out a solution o1 style. An evaluation system then rates the answer - strengthening successful reasoning patterns while weakening incorrect ones.

Example of RFT medical training: case description, instructions and correct answer (Gen FOXE3).
An example of reinforcement learning in medicine: based on symptoms, the AI model should determine the most likely genetic causes and justify its answer. | Image: OpenAI
Flow chart: FOXE3 gene and additional genes are analyzed by the grader system, output score 0.7
The diagram illustrates the evaluation of the correct answer (FOXE3) by a scoring algorithm. The evaluation is intended to reinforce the "thinking" process toward the correct answer. | Image: OpenAI

OpenAI says this approach works especially well for specialized fields like law, finance, engineering, and insurance that need deep technical knowledge. As an example, the company highlights its collaboration with Thomson Reuters, where they trained the compact o1 Mini model to work as a legal assistant.

Reinforcement learning for expert systems

Justin Ree, a bioinformatician at Berkeley Lab, used RFT to study rare genetic diseases. He trained the system using data extracted from hundreds of scientific papers that included symptoms and their associated genes.

Ad
Ad

Ree reports that the RFT-trained o1 Mini outperformed the standard o1 model at this task, despite being smaller and less expensive. He notes that the model's ability to explain its predictions makes it particularly useful.

Testing shows the fine-tuned mini model achieves the highest precision in gene identification, reaching up to 45 percent accuracy at maximum range.

Line chart: Comparison of gene identification accuracy for three model variants using different metrics
The fine-tuned mini model (01-mini finetune) achieves the highest precision in gene identification with up to 45 percent at maximum range. | Image: OpenAI

Early access program

OpenAI is now accepting organizations into its Reinforcement Fine-Tuning Research Program. The program is designed for organizations working on complex tasks that could benefit from AI assistance.

Participants will receive access to the RFT API and can help improve it through feedback before its public release. OpenAI plans to make RFT more widely available in early 2025.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • OpenAI has introduced a new training method for its o1 AI models called Reinforcement-Fine-Tuning (RFT), which goes beyond the previous supervised fine-tuning approach.
  • With RFT, the model is presented with a problem, given time to come up with a solution, and then the answer is evaluated, reinforcing successful reasoning and weakening incorrect ones.
  • OpenAI says RFT is particularly well suited to domains requiring deep domain knowledge, such as law, finance, engineering, and insurance, because it allows the model to learn the correct solution paths relevant to the task at hand.
Sources
Online journalist Matthias is the co-founder and publisher of THE DECODER. He believes that artificial intelligence will fundamentally change the relationship between humans and computers.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.