ollama it ??
which will better gguf quantized ?
No, it's not ollama/llamacpp, it's transformers.
I'm assuming the gguf F32 would be better...
· Sign up or log in to comment