No description
  • Go 96.4%
  • Makefile 1.9%
  • Dockerfile 1.2%
  • HCL 0.5%
Find a file
Dorin Geman b16c59b3d0 docs: archive notice - repository consolidated into model-runner
This repo has been consolidated into model-runner. All future development, issues, and PRs should be directed there.

Signed-off-by: Dorin Geman <dorin.geman@docker.com>
2025-10-06 12:21:19 +03:00
.github/workflows build: run on macos-latest 2025-09-23 09:05:41 +03:00
commands Merge pull request #76 from doringeman/ports-docs 2025-10-02 12:28:57 +03:00
desktop feat(run/chat): display token usage 2025-09-19 12:47:03 +03:00
desktop-module Fix path of the binary inside the module image 2025-03-19 10:24:59 +01:00
docs Merge pull request #76 from doringeman/ports-docs 2025-10-02 12:28:57 +03:00
mocks status: Add mock and UTs 2025-04-07 19:29:13 +03:00
pkg Merge pull request #76 from doringeman/ports-docs 2025-10-02 12:28:57 +03:00
vendor Merge pull request #153 from doringeman/render 2025-09-19 14:55:21 +01:00
.gitignore gitignore: Add dist/ 2025-03-18 11:48:55 +02:00
docker-bake.hcl ci: add a validate-tests bake target (#145) 2025-08-27 11:56:01 +01:00
Dockerfile ci: add a validate-tests bake target (#145) 2025-08-27 11:56:01 +01:00
go.mod Merge pull request #153 from doringeman/render 2025-09-19 14:55:21 +01:00
go.sum Merge pull request #153 from doringeman/render 2025-09-19 14:55:21 +01:00
LICENSE Add Apache2 License (#9) 2025-03-28 17:54:06 +01:00
main.go [AIE-186] Add detection of Model Runner context 2025-05-12 09:49:55 -06:00
Makefile build: set macOS 14 minimum version 2025-09-23 09:05:38 +03:00
README.md docs: archive notice - repository consolidated into model-runner 2025-10-06 12:21:19 +03:00

Docker Model CLI

⚠️ This Repository Has Moved

This repository has been consolidated into model-runner.

All future development, issues, and pull requests should be directed there.

Please visit the new repository for the latest updates and to contribute.


A powerful command-line interface for managing, running, packaging, and deploying AI/ML models using Docker. This CLI lets you install and control the Docker Model Runner, interact with models, manage model artifacts, and integrate with OpenAI and other backends—all from your terminal.

Features

  • Install Model Runner: Easily set up the Docker Model Runner for local or cloud environments with GPU support.
  • Run Models: Execute models with prompts or in interactive chat mode, supporting multiline input and OpenAI-style backends.
  • List Models: View all models available locally or via OpenAI, with options for JSON and quiet output.
  • Package Models: Convert GGUF files into Docker model OCI artifacts and push them to registries, including license and context size options.
  • Configure Models: Set runtime flags and context sizes for models.
  • Logs & Status: Stream logs and check the status of the Model Runner and individual models.
  • Tag, Pull, Push, Remove, Unload: Full lifecycle management for model artifacts.
  • Compose & Desktop Integration: Advanced orchestration and desktop support for model backends.

Building

  1. Clone the repo:
    git clone https://github.com/docker/model-cli.git
    cd model-cli
    
  2. Build the CLI:
    make build
    
  3. Install Model Runner:
    ./model install-runner
    
    Use --gpu cuda for GPU support, or --gpu auto for automatic detection.

Usage

Run ./model --help to see all commands and options.

Common Commands

  • model install-runner — Install the Docker Model Runner
  • model run MODEL [PROMPT] — Run a model with a prompt or enter chat mode
  • model list — List available models
  • model package --gguf <path> --push <target> — Package and push a model
  • model logs — View logs
  • model status — Check runner status
  • model configure MODEL [flags] — Configure model runtime
  • model unload MODEL — Unload a model
  • model tag SOURCE TARGET — Tag a model
  • model pull MODEL — Pull a model
  • model push MODEL — Push a model
  • model rm MODEL — Remove a model

Example: Interactive Chat

./model run llama.cpp "What is the capital of France?"

Or enter chat mode:

./model run llama.cpp
Interactive chat mode started. Type '/bye' to exit.
> """
Tell me a joke.
"""

Advanced

  • Packaging: Add licenses and set context size when packaging models for distribution.

Development

  • Run unit tests:
    make unit-tests
    
  • Generate docs:
    make docs
    

License

Apache 2.0