I'm also working on getting koboldcpp working with GPU support, currently it works for ingestion but for some reason the generation itself is still pretty slow.. will post an update when it's working well!
this post was submitted on 17 Jun 2023
19 points (100.0% liked)
LocalLLaMA
2593 readers
4 users here now
Community to discuss about LLaMA, the large language model created by Meta AI.
This is intended to be a replacement for r/LocalLLaMA on Reddit.
founded 2 years ago
MODERATORS