* [yzma v1.11.0](https://github.com/hybridgroup/yzma) – Go-based library for hardware-accelerated local inference with llama.cpp integration. * [Cactus React Native v1.10.4](https://github.com/cactus-compute/cactus-react-native) – React Native package for running AI language models locally with support for text, vision, and tool calling. * [llama-swap v198](https://github.com/mostlygeek/llama-swap) – Reliable on-demand model switching between local OpenAI-compatible inference servers without restarting applications.