A minimal local-first python/rust client for llama.cpp/kobold.cpp that provides persistent chat memory and real-time token streaming over a remote inference server.
-
Updated
Apr 28, 2026 - Python
A minimal local-first python/rust client for llama.cpp/kobold.cpp that provides persistent chat memory and real-time token streaming over a remote inference server.
Add a description, image, and links to the rkvw topic page so that developers can more easily learn about it.
To associate your repository with the rkvw topic, visit your repo's landing page and select "manage topics."