cmd ai models - nself-org/cli GitHub Wiki
List, pull, and remove local Ollama models with hardware compatibility markers.
nself ai local models <subcommand> [flags]
nself ai local models manages the set of Ollama models on the host. It cross-references the bundled model registry against detected hardware (RAM, VRAM, GPU) and marks each model as supported, marginal, or insufficient for the current machine.
The registry is bundled with the AI plugin binary and never fetched at runtime. Hardware detection works on macOS, Linux, and Windows; Apple Silicon unified memory is handled as a special case (1.5ร headroom threshold).
Lists all models from the registry alongside their hardware compatibility markers.
nself ai local models list [flags]
Output columns
| Column | Description |
|---|---|
| Model | Model identifier (e.g. llama3.1:8b) |
| Provider |
ollama / anthropic / google / openai
|
| Context | Context window in tokens |
| RAM | Minimum RAM required (GB) |
| Compat | โ supported ยท ๐ก marginal ยท ๐ insufficient ยท โ cloud |
| Installed | Whether the model is installed in Ollama |
Flags
| Flag | Default | Description |
|---|---|---|
--installed |
false | Show only models already installed in Ollama |
--json |
false | Output as JSON array |
Example
$ nself ai local models list
MODEL PROVIDER CONTEXT RAM COMPAT INSTALLED
llama3.1:8b ollama 128k 8 GB โ
yes
llama3.1:70b ollama 128k 40 GB ๐ no
mistral-nemo ollama 128k 12 GB ๐ก no
qwen2.5-coder:7b ollama 128k 6 GB โ
yes
claude-opus-4-7 anthropic 200k โ โ โ
claude-sonnet-4-6 anthropic 200k โ โ โ
gemini-2.5-pro google 1M โ โ โPulls a model from the Ollama library (streaming progress) and registers it with the AI plugin.
nself ai local models add <model> [flags]
Flags
| Flag | Default | Description |
|---|---|---|
--json |
false | Output progress events as JSON |
Example
$ nself ai local models add mistral-nemo
Pulling mistral-nemo...
โโโโโโโโโโโโโโโโโโโโโโโ 84% 3.1 GB / 3.7 GBRemoves a model from Ollama (with confirmation) and unregisters it from the AI plugin.
nself ai local models remove <model> [flags]
Flags
| Flag | Default | Description |
|---|---|---|
--yes |
false | Skip confirmation prompt |
--json |
false | Output as JSON |
Example
$ nself ai local models remove mistral-nemo
Remove mistral-nemo from Ollama? [y/N] y
Removed.Prints the models that the registry recommends for the detected hardware profile.
nself ai local models recommend [flags]
Example
$ nself ai local models recommend
Detected: Apple Silicon, 32 GB unified memory, no discrete GPU
Recommended models:
โ
llama3.1:8b (needs 8 GB โ comfortable)
โ
qwen2.5-coder:7b (needs 6 GB โ comfortable)
๐ก mistral-nemo (needs 12 GB โ marginal, may page)
๐ llama3.1:70b (needs 40 GB โ insufficient)| Mark | Meaning |
|---|---|
| โ | Host RAM/VRAM comfortably meets requirements |
| ๐ก | Host meets requirements marginally; model may page to swap |
| ๐ | Insufficient RAM or VRAM to run the model |
| โ | Cloud-only provider; no hardware requirement |
Apple Silicon: unified memory is counted toward both RAM and VRAM with a 1.5ร comfort threshold applied before emitting a โ .
The registry (models/registry.yml in the AI plugin) covers:
| Provider | Models |
|---|---|
| Anthropic | claude-opus-4-7, claude-sonnet-4-6, claude-haiku-4-5 |
| gemini-2.5-pro, gemini-2.5-flash, gemini-2.5-flash-lite | |
| OpenAI | gpt-4o, gpt-4o-mini |
| Ollama | llama3.1:8b, llama3.1:70b, mistral-nemo, qwen2.5-coder:7b |
The registry ships with the plugin binary. It is never fetched at runtime, so nself ai local models list works offline for metadata. Pulling a model requires network access.
-
nself ai local install, Install Ollama runtime -
nself ai local health, Check Ollama and plugin health -
plugin-ai, AI plugin reference