Documentation
How LLMFit thinks about local model fit.
LLMFit is a hardware-aware advisor. It does not replace inference runtimes like Ollama or llama.cpp. It decides which models, quantizations, and run paths make sense for a machine before you commit to a download or deployment.
Core model
What the engine evaluates.
Hardware shape
Total RAM, available RAM, CPU cores, GPU count, VRAM, backend, unified memory, and provider installation state.
Model shape
Parameter count, context length, release date, MoE behavior, quantization options, and provider availability.
Fit dimensions
Fit, speed, context, and quality scores are combined into a recommendation ranking rather than relying on a single benchmark.
Run path
GPU, CPU offload, CPU-only, and unified memory scenarios are treated differently because practical usability changes fast across them.
Fit levels
Operational labels, not marketing labels.
Perfect
The model fits comfortably and is positioned as a high-confidence local choice for the current hardware.
Good
Runnable with sensible resource usage. Good is often the right production default for conservative placement.
Marginal
Technically runnable, but with tighter memory pressure or a weaker expected throughput profile.
Too tight
Included for analysis, but usually excluded from top scheduling endpoints because it should not be treated as a safe deployment target.
Interfaces
One engine, multiple operational surfaces.
| Surface | Best for | Typical entrypoint |
|---|---|---|
| TUI | Interactive exploration, filtering, model comparison, plan mode | llmfit |
| CLI | Scripting, JSON export, repeatable audits, batch checks | llmfit recommend --json |
| REST API | Schedulers, agent integrations, internal platforms, node-local services | llmfit serve --host 0.0.0.0 --port 8787 |
| Desktop | macOS users who want the same logic with a graphical wrapper | cargo tauri build |
Runtime coverage
What LLMFit works with.
Repository layout
What ships in this codebase.
llmfit-core/
Hardware detection, model catalogs, fit scoring, planning logic, and provider integration.
llmfit-tui/
The terminal UI, classic CLI commands, and REST API entrypoints.
llmfit-desktop/
The Tauri-based macOS desktop wrapper for users who want the same engine with a GUI shell.
site/
The product site and docs property that can be served independently from the runtime tooling.
Next step