Stars
Open-core workflow engine powering Bubble Lab — and fully runnable, hostable, and extensible on its own.
Bootable Llamafile inference server with model weights built-in !!!! Exparimental Not BOOTABLE YET !!!
Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc
LostRuins / koboldcpp
Forked from ggml-org/llama.cppRun GGUF models easily with a KoboldAI UI. One File. Zero Install.