TinyLlama 1.1B Compact 1.1B Llama-compatible model; popular GGUF quantizations make it fast on CPUs/GPUs locally. 0500 Open-source Models# 1.1B# GGUF# Llama-compatible