TGArchive
·2 хв читання · 246 слів·👁 18.9K17

🦭 SEAL: The AI That Generates Its Own Training Data

MIT researchers created SEAL (Self-Adapting Language Models), an AI system that can improve itself by generating new training data on its own.

The idea is to move away from training models on "raw" text and instead give the model the ability to process that data for better retention and application. The developers compare this framework to a student: rather than cramming lectures, it's more effective to take notes, create diagrams, and add margin remarks.

How does this work?

1️⃣ The model generates self-edits based on the new context it gets. The original dataset is reorganized, and training parameters and instructions are created.

2️⃣ This instruction is then used to fine-tune the model before testing it. If self-edit improves performance, it is employed in subsequent training cycles to generate new instructions

3️⃣ Through this trial-and-error process, the model not only gets "smarter" but also more efficient in improving itself.

This idea was tested on tasks requiring abstract reasoning. The base model, Llama-3.2-1B, which initially couldn't solve a single task, achieved 72.5% success using the SEAL method.

💡 The significance of such methods stems not just from boosting the capacity of AI models but also from making better use of training data, which is already in limited supply. Epoch AI analysts estimate that between 2026 and 2032, we will run out of data to train AI. Systems like SEAL could be part of the solution.

#news #science @hiaimediaen

Відкрити в Telegram
Повернутись до каналу