#
llamacpp
Here are 52 public repositories matching this topic...
Kernels & AI inference engine for phone chips
-
Updated
Oct 1, 2025 - C++
Yet Another Language Model: LLM inference in C++/CUDA, no libraries except for I/O
-
Updated
Sep 13, 2025 - C++
Inference Vision Transformer (ViT) in plain C/C++ with ggml
c cpu ai computer-vision cpp image-classification edge-computing vision-transformer whisper-cpp llamacpp ggml
-
Updated
Apr 11, 2024 - C++
Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).
bloom falcon moe gemma mistral mixture-of-experts model-quantization multi-gpu-inference m2m100 llamacpp llm-inference internlm llama2 qwen baichuan2 mixtral phi-2 deepseek minicpm
-
Updated
Mar 15, 2024 - C++
LLM in Godot
-
Updated
Jun 23, 2024 - C++
Local LLMs in your DAW!
-
Updated
Sep 25, 2024 - C++
Getting an LLM to work with Godot.
-
Updated
Oct 11, 2023 - C++
Improve this page
Add a description, image, and links to the llamacpp topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the llamacpp topic, visit your repo's landing page and select "manage topics."