Define artificial intelligence models in Docker Compose applications

Docker Compose now allows you to declare artificial intelligence model dependencies directly in your application´s configuration, making model provisioning and life-cycle management seamless across local and cloud platforms.

Docker Compose enables developers to define artificial intelligence models as first-class components within their application stacks. By introducing the models top-level element in the Compose file, teams can specify model dependencies alongside traditional service definitions. This architecture ensures that the required models are provisioned and managed in tandem with application services, which streamlines deployment and fosters portability across various environments supporting the Compose specification, such as Docker Model Runner and certain cloud providers.

Prerequisites for leveraging this feature include Docker Compose version 2.38 or later and a platform compatible with Compose models. Compose models are defined under the models section, allowing developers to reference model images, set configuration parameters such as context size, and pass runtime flags to the inference engine. Services can bind to models via short or long syntax. The short syntax facilitates automatic generation of environment variables (e.g., LLM_URL), while the long syntax offers custom variable naming for additional flexibility. Examples provided cover declaring language models and embedding models, as well as specifying cloud-optimized configurations using labels.

Platform portability is a defining aspect of Compose models. When running locally with Docker Model Runner, the defined models are pulled, executed, and the appropriate environment variables are injected into containers. On cloud platforms that support Compose models, the same configuration file can trigger the use of managed artificial intelligence services, benefit from cloud-specific scaling, and utilize enhanced monitoring, logging, and model versioning. This approach supports seamless migration between local and cloud environments and helps teams manage model access, configuration, and updates consistently. Further documentation is available for deep dives into syntax, advanced options, and integration best practices.

65

Impact Score

Looking ahead at artificial intelligence and work in 2026

MIT Sloan researchers expect 2026 to bring a widening performance gap between humans and large language models, a push to scale responsible artificial intelligence deployments, and new questions about creativity, safety, and data access in the workplace.

Model autophagy disorder and the risk of self consuming Artificial Intelligence models

Glow New Media director Phil Blything warns that as Artificial Intelligence systems generate more online text, future language models risk training on their own synthetic output and degrading in quality. He draws a parallel with the early human driven web, arguing that machine generated content could undermine the foundations that made resources like Wikipedia possible.

Artificial intelligence and the new great divergence

A White House research paper compares the potential impact of artificial intelligence to the Industrial Revolution and examines whether it could trigger a new great divergence among nations. The report outlines how the Trump administration aims to secure American leadership through accelerated innovation, infrastructure, and deregulation.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.