Teaching large language models how to absorb new knowledge

Researchers at MIT have developed a self-adapting framework that lets large language models permanently internalize new information by generating and learning from their own self-edits. The method could help Artificial Intelligence agents update between conversations and adapt to changing tasks.

Large language models typically remain static after deployment: their learned weights do not persistently change when a user provides new information, and in-context learning is ephemeral across conversations. Researchers at MIT propose a different approach that leverages a model’s in-context strengths to teach it how to update its own weights. Their framework, called seal for self-adapting LLMs, prompts a model to create multiple rewrites of an input passage as synthetic training examples, analogous to students making study sheets from a lecture.

The model generates several candidate self-edits, quizzes itself on each one, and uses a trial-and-error method based on reinforcement learning to select the self-edit that most improves downstream performance. It then applies that chosen self-edit to update its internal weights so the information is permanently internalized. The framework also lets the model choose aspects of the learning process, including which synthetic data to use, the learning rate, and how many training iterations to run, effectively enabling the model to configure its own optimization strategy.

In evaluations the approach improved accuracy on question-answering tasks by nearly 15 percent and increased success rates on some skill-learning tasks by more than 50 percent, and a small model using the technique outperformed much larger large language models on several benchmarks. The researchers note an outstanding limitation: catastrophic forgetting, where performance on earlier tasks degrades as the model continually adapts. The paper’s authors include jyothish pari, adam zweiger, han guo, ekin akyürek, yoon kim, and pulkit agrawal, and the work will be presented at the conference on neural information processing systems. The team plans to address catastrophic forgetting and explore multi-agent settings, with support from the U.S. Army Research Office, the U.S. Air Force Artificial Intelligence Accelerator, the Stevens Fund for MIT UROP, and the MIT-IBM Watson Artificial Intelligence Lab.

63

Impact Score

Adobe plans outcome-based pricing for Artificial Intelligence agents

Adobe is positioning its Artificial Intelligence agents around performance-based pricing, charging only when the software completes useful work. The approach points to a more results-oriented model for selling generative Artificial Intelligence tools to business customers.

Tech firms commit billions to Artificial Intelligence infrastructure

Amazon, OpenAI, Nvidia, Meta, Google and others are signing increasingly large cloud, chip and data center agreements as demand for Artificial Intelligence infrastructure accelerates. The latest wave of deals spans investments, compute purchases, chip supply agreements and data center buildouts.

JEDEC outlines LPDDR6 expansion for data centers

JEDEC has previewed planned updates to LPDDR6 aimed at pushing the memory standard beyond mobile devices and into selected data center and accelerated computing use cases. The roadmap includes higher-capacity packaging options, flexible metadata support, 512 GB densities, and a new SOCAMM2 module standard.

Tsmc debuts A13 process technology

Tsmc has introduced its A13 process at its 2026 North America Technology Symposium as a tighter version of A14 aimed at next-generation Artificial Intelligence, high performance computing, and mobile designs. The company positions the node as a more compact and efficient option with backward-compatible design rules for faster migration.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.