llama.cpp logo

llama.cpp

Repo

llama.cpp is an open-source C/C++ implementation for running quantized Llama and other GGUF large-language-model weights locally on CPUs and GPUs. It targets developers who want lightweight, offline inference without heavy framework dependencies.

article 2 storys calendar_today First: 2026-03-04 update Last: 2026-04-02 menu_book Wikipedia

Stories

Completed digest stories linked to this service.

GET_DAILY_EMAIL
AI + SDLC // 5 MIN DAILY