Teaching large language models how to absorb new knowledge

Researchers at MIT have developed a self-adapting framework that lets large language models permanently internalize new information by generating and learning from their own self-edits. The method could help Artificial Intelligence agents update between conversations and adapt to changing tasks.

NVIDIA sweeps MLPerf training v5.1 for artificial intelligence

NVIDIA swept all seven tests in MLPerf Training v5.1, posting the fastest training times across large language models, image generation, recommender systems, computer vision and graph neural networks. The company was the only platform to submit results on every test, highlighting its GPUs and CUDA software stack.