IBM Unveils Granite 4.0 Tiny Preview: A Compact, Efficient Language Model

IBM releases the Granite 4.0 Tiny Preview, a compact and compute-efficient Artificial Intelligence model, bringing advanced language processing to consumer hardware.

IBM has introduced the Granite 4.0 Tiny Preview, the initial release of its most compact model in the forthcoming Granite 4.0 language model family, targeting the open source community. Emphasizing both minimal memory requirements and high efficiency, Granite 4.0 Tiny Preview leverages FP8 precision to allow multiple, long-context (128,000 token) sessions to be executed concurrently on widely accessible consumer GPUs. Despite being only partially trained—processing about 2.5 trillion of an expected 15 trillion training tokens—the model already achieves performance competitive with IBM´s Granite 3.3 2B Instruct model, while utilizing significantly fewer parameters and offering approximately 72% less memory usage.

Set to become the smallest offering within the Granite 4.0 family, the Tiny Preview is just the start of a suite that will also include Small and Medium variants, all slated for official release later this summer. The current Granite 4.0 iteration reinforces IBM’s commitment to accessible and efficient enterprise large language models (LLMs), making high-performance Artificial Intelligence more attainable for a broad user base. The preview model, distributed on Hugging Face under the open-source Apache 2.0 license, is not yet recommended for enterprise deployment, but it is designed for experimentation by developers—even those with limited hardware resources.

The architecture introduced in Granite 4.0 Tiny is still awaiting integration with key platforms such as Hugging Face Transformers and vLLM, with support expected in the near future. Additionally, IBM plans to enable easy local deployment of Granite 4.0 models through partners like Ollama and LMStudio following the official release. As the training of Granite 4.0 Tiny progresses, IBM anticipates its final performance will match that of larger predecessors, such as the Granite 3.3 8B Instruct, further closing the gap between compactness and robust language capabilities. This approach underscores IBM´s strategic focus on practical, resource-conscious Artificial Intelligence innovation for both research and enterprise environments.

66

Impact Score

Research excellence at the UF College of Medicine in 2025

In 2025, the University of Florida College of Medicine expanded its research footprint across cancer, neuromedicine, diabetes, and women’s and children’s health, leveraging artificial intelligence to accelerate discovery and clinical impact.

What EO 14365 means for state artificial intelligence laws and business compliance

Executive order 14365 signals a push toward a national artificial intelligence policy that could preempt certain state regulations without immediately changing existing compliance obligations. Businesses using artificial intelligence are advised to monitor forthcoming federal actions while continuing to follow current state laws.

Generative Artificial Intelligence reshapes europe’s economy, society and policy

The european commission’s joint research centre outlines how generative artificial intelligence is altering research, industry, labour markets and social equality in the EU, while highlighting gaps in patents, investment and safeguards. The report points to both productivity gains and rising risks that demand coordinated policy responses.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.