Training artificial intelligence models to assimilate new knowledge

MIT researchers developed a method that lets large language models generate study-like synthetic data and update their weights to permanently internalize new information, a step toward self-improving artificial intelligence.

Large language models deployed today cannot permanently learn from new information the way humans do. The article explains that once a model’s training is complete its internal weights remain static, so information provided during a conversation does not persist across sessions. Although models perform well at in-context learning, where they use examples within a single interaction to guide responses, that knowledge disappears when the session ends.

Researchers at MIT introduced a framework called SEAL for “self-adapting LLMs” that aims to teach models how to update their own weights using synthetic training data. The model rewrites incoming information into multiple self-edits, akin to a student creating study sheets, and then evaluates each self-edit by quizzing itself on downstream tasks. Using a reinforcement learning approach, the model rewards the self-edits that produce the largest performance gains and then applies the best edit to its weights so the knowledge is internalized.

In experiments the SEAL method improved accuracy on question-answering tasks by nearly 15 percent and boosted success rates on some skill-learning tasks by more than 50 percent, with a small model outperforming much larger models on certain benchmarks. The authors note a key limitation: catastrophic forgetting, where adapting to new information can erode prior knowledge. The team plans further work to mitigate forgetting and to explore multi-agent settings in which models teach each other. The research was led by MIT students and faculty and will be presented at the Conference on Neural Information Processing Systems, with support from several funding agencies including the U.S. Army Research Office and the U.S. Air Force AI Accelerator.

68

Impact Score

Global regulatory trends on the use of generative artificial intelligence

Governments in the EU, Japan, the United States, and the United Kingdom are moving quickly to regulate generative artificial intelligence, using a mix of binding laws, guidelines, and standards. Diverging philosophies and timelines are making cross-border compliance planning increasingly complex for companies.

Perplexity launches Computer to orchestrate many Artificial Intelligence models

Perplexity is rolling out Computer, a cloud-based agent that coordinates 19 Artificial Intelligence models for complex workflows, as it pivots toward high-value enterprise users and deep research. The launch underscores a broader bet on multi-model orchestration, custom benchmarks and a boutique business strategy over mass adoption.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.