DeepSeek explores image tokens to improve Artificial Intelligence memory

DeepSeek released an optical character recognition model that packs text into images to store more context with fewer tokens, a technique the company says could improve Artificial Intelligence memory and efficiency.

DeepSeek released an optical character recognition model last week that extracts text from images and turns it into machine-readable words. The paper and early reviews report that the model performs on par with top systems on key benchmarks for optical character recognition. DeepSeek presents the model primarily as a test bed for a different approach to storing and retrieving information inside Artificial Intelligence systems.

The paper’s central innovation is replacing large sets of text tokens with image-based tokens and a tiered compression scheme. Instead of representing words as thousands of discrete tokens, the system packs written information into image form, similar to photographing pages of a book, and stores older or less critical content in progressively blurrier representations. The company argues this lets models retain nearly the same information using far fewer tokens, which could cut the computing resources required to run long conversations and reduce the carbon footprint associated with Artificial Intelligence memory. The paper frames the method as a possible remedy for so-called context rot, where long interactions cause models to forget or muddle earlier information.

Researchers have already noted the approach’s promise while cautioning that it is an early exploration. Andrej Karpathy praised the paper on X, suggesting images may be better than text as inputs for large language models and criticizing text tokens as wasteful. Manling Li of Northwestern University called the work a new framework for addressing memory challenges, and Zihan Wang highlighted its potential to help continuous conversations remember more effectively. DeepSeek also reports that the system can generate more than 200,000 pages of training data a day on a single GPU. Based in Hangzhou, DeepSeek previously surprised the industry with DeepSeek-R1 earlier this year, and the new paper continues its push into low-resource research directions while noting more work is needed to make memory recall more dynamic and importance-aware.

53

Impact Score

Vals publishes public enterprise language model benchmarks

Vals lists a broad set of public enterprise benchmarks spanning law, finance, healthcare, math, education, academics, coding, and beta agent tasks. The index highlights which models currently lead specific enterprise-focused evaluations and how widely each benchmark has been tested.

MIT method spots overconfident Artificial Intelligence models

MIT researchers developed a way to detect when large language models are confidently wrong by comparing their answers with outputs from similar models. The combined uncertainty measure outperformed standard techniques across a range of tasks and may help reduce unreliable responses.

MEPs back delay for parts of Artificial Intelligence Act

European Parliament committees have endorsed targeted delays to parts of the Artificial Intelligence Act while adding a proposed ban on certain non-consensual image manipulation tools. The changes aim to give companies clearer deadlines, reduce overlap with other EU rules, and extend support to small mid-cap enterprises.

Publisher alliance seeks leverage over Artificial Intelligence web access

A new publisher coalition is trying to reshape how Artificial Intelligence companies access journalism by combining collective bargaining with tougher technical controls. The effort reflects growing pressure on Artificial Intelligence firms to pay for content used in training, search, and user-facing responses.

Military advantage in the age of algorithmic diffusion

American leadership in Artificial Intelligence research and infrastructure may not translate into lasting military advantage. Rapid diffusion of algorithms is shifting the contest toward compute, talent, and the speed of military adoption.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.