-
Notifications
You must be signed in to change notification settings - Fork 2.9k
Description
Problem (one or two sentences)
I try to use Inferencer app with RooCode https://inferencer.com/
Inferencer seems to work noticeable faster than Ollama with local models. I have tested it with GLM-4.7-Flash, so I say it after working with this model exactly. It also supports MLX models. As I had problems with LM Studio due to more memory it requires, this is the only MLX supporting runner working for me for now.
Context (who is affected and when)
RooCode v3.21.3
Inferencer v.1.9.5
Inferencer has option to provide "Ollama compatible API", and this is JSON response of its /api/tags endpoint at my machine:
{
"models": [
{
"name": "/inferencerlabs/glm47flashmlx:o",
"model": "/inferencerlabs/glm47flashmlx:o",
"modified_at": "2026-02-08T23:57:04.212606191+01:00",
"size": 24336701885,
"digest": "54515c0a8d28bed854515c0a8d28bed854515c0a8d28bed854515c0a8d28bed8",
"details": {}
},
{
"name": "/blobs/inferencerlabs/glm47flashmlx:o",
"model": "/blobs/inferencerlabs/glm47flashmlx:o",
"modified_at": "2026-02-08T23:57:04.212606191+01:00",
"size": 24336701885,
"digest": "64e5b35b68419b2164e5b35b68419b2164e5b35b68419b2164e5b35b68419b21",
"details": {}
}
]
}
As you see it lacks "details" key. I think this is the problem which causes RooCode not to find any of the provided models.
However, I have managed to run Inferencer with Roo Code v3.21.2, and it works like a charm :D. Then v3.21.3 already does not detect models from Inferencer. I tested multiple newer versions including the newest 3.47.3 and non of them was able to connect to the model.
What I see at RooCode v3.21.3 and newer:
The model ID (/inferencerlabs/glm47flashmlx:o) you provided is not available. Please choose a different model.
Desired behavior (conceptual, not technical)
Possibility to show all models provided by Ollama API even if only minimum information about the model like at the attached JSON is available. Possibility to connect with any of these models.
Constraints / preferences (optional)
No response
Request checklist
- I've searched existing Issues and Discussions for duplicates
- This describes a specific problem with clear context and impact
Roo Code Task Links (optional)
No response
Acceptance criteria (optional)
No response
Proposed approach (optional)
I understand that there are probably more Ollama API features used at "Ollama" model provider, and they are needed. The proposition is to add "Ollama compatible" mode with minimal API requirement at least when it comes to models discovery. It would be handy for cooperating with third party tools like Inferencer.
Trade-offs / risks (optional)
No response