I was wondering if adding Ollama as a backend server would be an option. It's a polished wrapper for llama cpp I believe. https://ollama.com
I was wondering if adding Ollama as a backend server would be an option. It's a polished wrapper for llama cpp I believe. https://ollama.com