Skip to content

[ENHANCEMENT] "Ollama compatible" model provider #11339

@kneczaj

Description

@kneczaj

Problem (one or two sentences)

I try to use Inferencer app with RooCode https://inferencer.com/

Inferencer seems to work noticeable faster than Ollama with local models. I have tested it with GLM-4.7-Flash, so I say it after working with this model exactly. It also supports MLX models. As I had problems with LM Studio due to more memory it requires, this is the only MLX supporting runner working for me for now.

Context (who is affected and when)

RooCode v3.21.3
Inferencer v.1.9.5

Inferencer has option to provide "Ollama compatible API", and this is JSON response of its /api/tags endpoint at my machine:

{
  "models": [
    {
      "name": "/inferencerlabs/glm47flashmlx:o",
      "model": "/inferencerlabs/glm47flashmlx:o",
      "modified_at": "2026-02-08T23:57:04.212606191+01:00",
      "size": 24336701885,
      "digest": "54515c0a8d28bed854515c0a8d28bed854515c0a8d28bed854515c0a8d28bed8",
      "details": {}
    },
    {
      "name": "/blobs/inferencerlabs/glm47flashmlx:o",
      "model": "/blobs/inferencerlabs/glm47flashmlx:o",
      "modified_at": "2026-02-08T23:57:04.212606191+01:00",
      "size": 24336701885,
      "digest": "64e5b35b68419b2164e5b35b68419b2164e5b35b68419b2164e5b35b68419b21",
      "details": {}
    }
  ]
}

As you see it lacks "details" key. I think this is the problem which causes RooCode not to find any of the provided models.

However, I have managed to run Inferencer with Roo Code v3.21.2, and it works like a charm :D. Then v3.21.3 already does not detect models from Inferencer. I tested multiple newer versions including the newest 3.47.3 and non of them was able to connect to the model.

What I see at RooCode v3.21.3 and newer:
The model ID (/inferencerlabs/glm47flashmlx:o) you provided is not available. Please choose a different model.

Desired behavior (conceptual, not technical)

Possibility to show all models provided by Ollama API even if only minimum information about the model like at the attached JSON is available. Possibility to connect with any of these models.

Constraints / preferences (optional)

No response

Request checklist

  • I've searched existing Issues and Discussions for duplicates
  • This describes a specific problem with clear context and impact

Roo Code Task Links (optional)

No response

Acceptance criteria (optional)

No response

Proposed approach (optional)

I understand that there are probably more Ollama API features used at "Ollama" model provider, and they are needed. The proposition is to add "Ollama compatible" mode with minimal API requirement at least when it comes to models discovery. It would be handy for cooperating with third party tools like Inferencer.

Trade-offs / risks (optional)

No response

Metadata

Metadata

Assignees

No one assigned

    Labels

    EnhancementNew feature or request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions