TUI and CLI for browsing AI models, benchmarks, and coding agents.
- Models Tab: Browse 3,000+ models across 85+ providers from models.dev with capability indicators, adaptive layouts, and provider categorization
- Benchmarks Tab: Compare model performance across 15+ benchmarks from Artificial Analysis, with head-to-head tables, scatter plots, radar charts, and creator filtering
- Agents Tab: Track AI coding assistants (Claude Code, Aider, Cursor, etc.) with version detection, changelogs, and GitHub integration
models-demo.mp4
Video (and screenshots below) are out-of-sync with the current state of the app, I've been moving fast on making changes and so I'll have to record a new one!
- Models tab redesign — capability indicators, adaptive provider panel, and detailed model info at a glance
- Benchmark compare mode — head-to-head tables, scatter plots, and radar charts for selected models
- Benchmarks CLI — list and inspect benchmark data directly from the terminal
- Linux packages — native .deb and .rpm packages for x86_64 and aarch64
- Agents CLI — track agent releases, view changelogs, and compare versions from the terminal
- Capability indicators — see Reasoning, Tools, Files, and Open/Closed status at a glance in the model list
- Provider categories — filter and group providers by type (Origin, Cloud, Inference, Gateway, Dev Tool)
- Detail panel — capabilities, pricing, modalities, and metadata for the selected model
- Cross-provider search to compare the same model across different providers
- Copy to clipboard with a single keypress
- CLI commands and JSON output for scripting and automation
- Curated catalog of 12+ AI coding assistants
- Version detection — automatically detects installed agents
- GitHub integration — stars, releases, changelogs, update availability
- Styled changelogs — markdown rendering with syntax highlighting in the detail pane
- Changelog search — search across changelogs with highlighted matches and
n/Njump-to-match - Persistent cache — instant startup with ETag-based conditional fetching
- Customizable tracking — choose which agents to monitor
- ~400 benchmark entries from Artificial Analysis with quality, speed, and pricing scores
- Compare mode — select models for head-to-head tables, scatter plots, and radar charts
- Auto-updating — benchmark data refreshed automatically every 30 minutes
- Creator sidebar with 40+ creators — filter by region, type, or open/closed source
- Sort & filter — sort by any metric, filter by reasoning capability, source type, and more
- Detail panel — full benchmark breakdown with indexes, scores, performance, and pricing
- Status table — see installed vs latest version, 24h release indicator, and release frequency at a glance
- Inline release browser —
agents <tool>opens an interactive version browser with changelog preview - Changelogs — view release notes for any agent by name, latest version, or explicit version
- Tracked-agent manager —
agents list-sourcescan now manage which curated agents are tracked from the CLI - Dual entry point — use as
models agentsor create anagentssymlink for standalone usage - Fast — concurrent GitHub fetching and version detection
- Live benchmark queries — fetch the current benchmark dataset without launching the TUI
- Interactive list picker — use
models benchmarks listto open a filtered benchmark selector, then inspect the selected model immediately - Detail views — use
models benchmarks showfor a direct model breakdown, with interactive disambiguation when a query matches multiple variants - Filtering — narrow by search text, creator, open/closed source, and reasoning status
- Sorting — sort by any supported metric, including intelligence, coding, math, GPQA, speed, pricing, and release date
- JSON output — pipe structured benchmark data into shell scripts and other tools
cargo install modelsdevbrew install modelsMigrating from the tap? Run
brew untap arimxyer/tap— updates now land throughhomebrew-corebump PRs and may take a bit to merge.
scoop install extras/modelsMigrating from the custom bucket? Run
scoop bucket rm arimxyer— Scoop Extras handles updates automatically.
paru -S models-bin # or: yay -S models-binMaintained by @Dominiquini
Download the .deb from GitHub Releases and install:
# Download the latest .deb for your architecture (amd64 or arm64)
sudo dpkg -i modelsdev_*_amd64.debDownload the .rpm from GitHub Releases and install:
# Download the latest .rpm for your architecture (x86_64 or aarch64)
sudo rpm -i modelsdev-*.x86_64.rpmVerifying downloads: Each GitHub Release includes a
SHA256SUMSfile. After downloading, verify with:sha256sum -c SHA256SUMS --ignore-missing
Download the latest release for your platform from GitHub Releases.
git clone https://github.com/arimxyer/models
cd models
cargo build --release
./target/release/modelsRun models with no arguments to launch the interactive TUI:
modelsGlobal
| Key | Action |
|---|---|
] / [ |
Switch tabs (Models / Agents / Benchmarks) |
? |
Show context-aware help |
q |
Quit |
Navigation
| Key | Action |
|---|---|
j / ↓ |
Move down |
k / ↑ |
Move up |
g |
Jump to first item |
G |
Jump to last item |
Ctrl+d / PageDown |
Page down |
Ctrl+u / PageUp |
Page up |
Tab / Shift+Tab |
Switch panels |
← / → |
Switch panels |
Search
| Key | Action |
|---|---|
/ |
Enter search mode |
Enter / Esc |
Exit search mode |
Esc |
Clear search (in normal mode) |
Filters & Sort
| Key | Action |
|---|---|
s |
Cycle sort (name → date → cost → context) |
S |
Toggle sort direction (asc/desc) |
1 |
Toggle reasoning filter |
2 |
Toggle tools filter |
3 |
Toggle open weights filter |
4 |
Toggle free models filter |
5 |
Cycle provider category filter (All → Origin → Cloud → Inference → Gateway → Tool) |
6 |
Toggle category grouping |
Copy & Open
| Key | Action |
|---|---|
c |
Copy provider/model-id |
C |
Copy model-id only |
o |
Open provider docs in browser |
D |
Copy provider docs URL |
A |
Copy provider API URL |
Filters & Sort
| Key | Action |
|---|---|
s |
Cycle sort (name → updated → stars → status) |
1 |
Toggle installed filter |
2 |
Toggle CLI tools filter |
3 |
Toggle open source filter |
Search
| Key | Action |
|---|---|
/ |
Search agents and changelogs |
n |
Jump to next match |
N |
Jump to previous match |
Actions
| Key | Action |
|---|---|
a |
Open tracked agents picker |
o |
Open docs in browser |
r |
Open GitHub repo |
c |
Copy agent name |
By default, models tracks 4 popular agents: Claude Code, Codex, Gemini CLI, and OpenCode.
Press a in the Agents tab to open the picker and customize which agents you track. Your preferences are saved to ~/.config/models/config.toml.
You can also add custom agents not in the catalog:
# ~/.config/models/config.toml
[[agents.custom]]
name = "My Agent"
repo = "owner/repo"
binary = "my-agent"
version_command = ["--version"]See Custom Agents for the full reference.
Quick Sort (press again to toggle direction)
| Key | Action |
|---|---|
1 |
Sort by Intelligence index |
2 |
Sort by Release date |
3 |
Sort by Speed (tok/s) |
Filters
| Key | Action |
|---|---|
4 |
Cycle source filter (All / Open / Closed) |
5 |
Cycle region filter (US / China / Europe / ...) |
6 |
Cycle type filter (Startup / Big Tech / Research) |
7 |
Cycle reasoning filter (All / Reasoning / Non-reasoning) |
Sort
| Key | Action |
|---|---|
s |
Open sort picker popup |
S |
Toggle sort direction (asc/desc) |
Compare Mode
| Key | Action |
|---|---|
Space |
Toggle model selection (max 8) |
v |
Cycle view (H2H table → Scatter → Radar) |
t |
Toggle left panel (Models / Creators) |
d |
Show detail overlay (H2H view) |
c |
Clear all selections |
h / l |
Switch focus (List / Compare) |
j / k |
Scroll H2H table (when Compare focused) |
x / y |
Cycle scatter plot axes |
a |
Cycle radar chart preset |
Actions
| Key | Action |
|---|---|
o |
Open Artificial Analysis page |
Query benchmark data from the command line using the same live benchmark feed as the Benchmarks tab.
models benchmarks list
models benchmarks list --sort speed --limit 10
models benchmarks list --creator openai --reasoning
models benchmarks list --open --sort price-input --ascmodels benchmarks list opens the inline picker in an interactive terminal and uses the same filters/sorting to narrow the candidate set before you pick a model.
Once the picker is open:
/starts a live text filter over name, slug, and creatorscycles sort metricsSreverses the current sortEnterprints the selected model's normalshowoutput
models benchmarks show gpt-4o
models benchmarks show "Claude Sonnet 4"If show matches multiple benchmark variants in an interactive terminal, the CLI reopens the picker with just the matching candidates so you can choose the exact row you want.
models benchmarks list --creator anthropic --json
models benchmarks show gpt-4o --jsonTrack AI coding agent releases from the command line. Install the agents alias during setup, or use models agents as a fallback.
# Create the agents alias (one-time setup)
mkdir -p ~/.local/bin
ln -s $(which models) ~/.local/bin/agentsNote: Make sure
~/.local/binis in yourPATH. For example, in bash/zsh addexport PATH="$HOME/.local/bin:$PATH"to your shell config, or in fish runfish_add_path ~/.local/bin.
agents status┌──────────────┬─────┬───────────┬──────────┬─────────┬───────────────┐
│ Tool │ 24h │ Installed │ Latest │ Updated │ Freq. │
├──────────────┼─────┼───────────┼──────────┼─────────┼───────────────┤
│ Claude Code │ ✓ │ 2.1.42 │ 2.1.42 │ 1d ago │ ~1d │
│ OpenAI Codex │ ✓ │ 0.92.0 │ 0.92.0 │ 6h ago │ ~3h │
│ Goose │ │ — │ 1.0.20 │ 3d ago │ ~2d │
└──────────────┴─────┴───────────┴──────────┴─────────┴───────────────┘
agents claude # Interactive release browser (by CLI binary name)
agents claude-code # By agent ID
agents claude --latest # Latest release directly
agents claude --version 1.0.170 # Specific versionagents claude --list # List all versions
agents claude --pick # Alias for the interactive release browserIn the release browser:
↑/↓orj/kmoves between releases- the lower pane previews the selected release notes
Enterprints the full changelog for the selected release
agents latest # Interactive picker for releases from the last 24 hours
agents list-sources # Interactive tracked-agent manager
agents claude --web # Open GitHub releases in browsermodels list
models list anthropicmodels list opens the inline picker in an interactive terminal. Use a provider argument to prefilter the picker before it opens.
Once the picker is open:
/starts a live filter over model id, name, and providerscycles sort modesSreverses the current sortEnterprints the selected model's normalshowoutput
models providers
models providers --jsonmodels show claude-opus-4-5-20251101Claude Opus 4.5
===============
ID: claude-opus-4-5-20251101
Provider: Anthropic (anthropic)
Family: claude-opus
Limits
------
Context: 200k tokens
Max Output: 64k tokens
Pricing (per million tokens)
----------------------------
Input: $5.00
Output: $25.00
Cache Read: $0.50
Cache Write: $6.25
Capabilities
------------
Reasoning: Yes
Tool Use: Yes
Attachments: Yes
Modalities: text, image, pdf -> text
Metadata
--------
Released: 2025-11-01
Updated: 2025-11-01
Knowledge: 2025-03-31
Open Weights: No
If show matches multiple providers or model variants in an interactive terminal, the CLI reopens the picker with the matching candidates so you can choose the exact row.
models search "gpt-4"
models search "claude opus"models search currently reuses the same matcher and interactive picker flow as models list, so it remains available as a compatibility command.
All models and benchmarks commands support --json for scripting:
models benchmarks list --json
models benchmarks show gpt-4o --json
models list --json
models providers --json
models show claude-opus-4-5 --json
models search "llama" --jsonLots of gratitude to the companies who do all the hard work! Shout out to the sources:
- Model data: Fetched from models.dev, an open-source database of AI models maintained by SST
- Benchmark data: Fetched from Artificial Analysis — quality indexes, benchmark scores, speed, and pricing for ~400 model entries
- Agent data: Curated catalog in
data/agents.json— contributions welcome! - GitHub data: Fetched from GitHub API (stars, releases, changelogs)
- Nix flake — Nix packaging with a proper
flake.lockfor reproducible builds (PRs welcome!)
Contributions are welcome! Please read the Contributing Guide before submitting a PR.
This project follows the Contributor Covenant Code of Conduct.
MIT


