Text Generation WebUI (oobabooga) All-in-one local web UI for LLMs with chat, memory, extensions and RP-friendly presets; runs GGUF/llama.cpp and more. 0570 Roleplay Frontends (Local/Self-Hosted UI)# adult# extensions# GGUF
KoboldCpp Single-binary llama.cpp fork with a lightweight built-in UI; fast local inference for RP with long context options. 0370 Roleplay Frontends (Local/Self-Hosted UI)# adult# GGUF# inference
TinyLlama 1.1B Compact 1.1B Llama-compatible model; popular GGUF quantizations make it fast on CPUs/GPUs locally. 0500 Open-source Models# 1.1B# GGUF# Llama-compatible
Ollama Library Pull open-weight LLMs locally with one command; browse popular chat and embedding models. 0420 Model Hubs# embeddings# GGUF# LLMs