graph TD;
ggml --> whisper.cpp
ggml --> llama.cpp
llama.cpp --> llama.vim
llama.cpp --> llama.vscode
ggml[<a href="https://github.com/ggml-org/ggml" style="text-decoration:none;">ggml</a>];
whisper.cpp[<a href="https://github.com/ggml-org/whisper.cpp" style="text-decoration:none;">whisper.cpp</a>];
llama.cpp[<a href="https://github.com/ggml-org/llama.cpp" style="text-decoration:none;">llama.cpp</a>];
llama.vim[<a href="https://github.com/ggml-org/llama.vim" style="text-decoration:none;">llama.vim</a>];
llama.vscode[<a href="https://github.com/ggml-org/llama.vscode" style="text-decoration:none;">llama.vscode</a>];
- Cloud: https://endpoints.huggingface.co (tutorial)
- Jobs: jobs@ggml.ai (full-time, only active contributors will be considered)
ggml
/llama.cpp
maintainer
- Business: sales@ggml.ai
[2025 Apr 18]
Gemma 3 QAT Models: Bringing state-of-the-Art AI to consumer GPUs[2025 Apr 16]
Llama 4 Runs on Arm[2025 Apr 04]
Run LLMs Locally with Docker[2025 Mar 25]
Deploy a Large Language Model (LLM) chatbot with llama.cpp using KleidiAI on Arm servers[2025 Feb 11]
OLMoE, meet iOS[2024 Oct 02]
Accelerating LLMs with llama.cpp on NVIDIA RTX Systems
Chat | STT | Mobile | Infra | Cloud |
---|---|---|---|---|
LM Studio | MacWhisper | PocketPal AI | RamaLama | Hugging Face |
KoboldCpp | VLC media player | LLMFarm | paddler | |
LocalAI | WASM | llama-swap | ||
Jan |