Hi everyone,
I just started experimenting with the LLM Application (v0.7.2). I’m not sure if what I’m experiencing is a misconfiguration on my end or a potential bug.
My Setup:
XWiki version: 17.7.0 (Debian)
LLM Application: 0.7.2
Only one provider configured: Ollama
Ollama server: Running on a separate machine in my local network
I added the provider via the AI LLM Application config in the administrator menu – just with a name and the correct URL, and streaming activated.
A query (curl http://*:11434/v1/models
) to my Ollama looks like this:
{“object”:“list”,“data”:[{“id”:“llama3.1:latest”,“object”:“model”,“created”:1756733917,“owned_by”:“library”},{“id”:“qwen2:latest”,“object”:“model”,“created”:1756733814,“owned_by”:“library”},{“id”:“llama3:latest”,“object”:“model”,“created”:1756733053,“owned_by”:“library”},{“id”:“zephyr:latest”,“object”:“model”,“created”:1756732566,“owned_by”:“library”},{“id”:“gemma:2b”,“object”:“model”,“created”:1756732469,“owned_by”:“library”}]}
However, when I check the XWiki internal API:
GET /xwiki/rest/wikis/xwiki/aiLLM/v1/models
…I get an empty result:
{“data”:[],“object”:“list”,“first_id”:null,“last_id”:null,“has_more”:false}
Even after restarting XWiki, nothing changes.
What does work:
If I manually add a model in the configuration UI (e.g., “llama3.1:latest”), then it shows up in the list, and everything works from there — completions, chat, etc.
What I’ve Checked:
-
/models and /chat/completions endpoints are reachable from the XWiki server
-
CORS is working correctly (tested via curl and browser)
-
No relevant errors in the logs
-
Restarted XWiki
Is this a known bug, or am I missing a step?
I’d really appreciate any pointers. Thanks in advance.