Ollama alternatives #28
Replies: 2 comments 1 reply
-
|
Hey @zyfer9009-lgtm, thanks for reaching out from Denmark! Appreciate the kind words. Good news - this is definitely possible since FastFlowLM exposes an OpenAI-compatible API. CAAL currently supports Ollama and Groq, but the architecture is set up to add new providers. I've created a feature request for this: #29 The Groq provider we added recently uses the same response format as OpenAI, so we can reuse most of that code - just need to point it at a configurable endpoint instead of Groq's cloud. Question: Would you be willing to test a feature branch when we start implementing this? Having someone with actual FastFlowLM hardware to validate against would be really helpful. No ETA yet, but it's on the radar. |
Beta Was this translation helpful? Give feedback.
-
|
Hey @zyfer9009-lgtm — good news! We just shipped v1.6.0 which adds an OpenAI-compatible provider. This should work with FastFlowLM out of the box. To set it up:
You can also set it in Let me know how it goes with your AMD NPU setup — would love to hear if it works well! |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi there.😊 Really great work you are doing👌. Just a question from Denmark. I have a setup where I use AMD NPU via FastflowLM, not GPU. Is there a way to connect to FastFlowLM, instead of Ollama. I can connect via Open-AI Api (Dummy). I'm currently using Open webui, and it works fine, but not good at voice commands. Hope you can help, because I really want to try your cool CAAL..
Beta Was this translation helpful? Give feedback.
All reactions