* [yzma v1.11.0](https://github.com/hybridgroup/yzma) – Go-based library for hardware-accelerated local inference with llama.cpp integration. * [Jan - Open-source ChatGPT replacement v0.7.8](https://github.com/janhq/jan) – Runs local and cloud AI models with privacy-focused control and customizable assistants. * [Cactus React Native v1.10.0](https://github.com/cactus-compute/cactus-react-native) – React Native package for running AI language models locally with support for text, vision, and tool calling. * [llama-swap v198](https://github.com/mostlygeek/llama-swap) – Reliable on-demand model switching between local OpenAI-compatible inference servers without restarting applications.