diff --git a/ollama-integration/info.json b/ollama-integration/info.json index cfe8320..5d4fe41 100644 --- a/ollama-integration/info.json +++ b/ollama-integration/info.json @@ -2,7 +2,7 @@ "name": "Ollama AI backend integration", "identifier": "ollama-integration", "script": "ollama-integration.qml", - "version": "0.1.2", + "version": "0.1.3", "minAppVersion": "24.6.3", "authors": ["@pbek"], "description" : "This script provides integration for a local Ollama AI backend. If the Ollama service is up, the installed models will be fetched automatically every time the scripting engine is reloaded." diff --git a/ollama-integration/ollama-integration.qml b/ollama-integration/ollama-integration.qml index a3a77ec..77466f8 100644 --- a/ollama-integration/ollama-integration.qml +++ b/ollama-integration/ollama-integration.qml @@ -30,7 +30,7 @@ Script { ]; function init() { - const data = script.downloadUrlToString(this.apiBaseUrl + '/api/tags') + const data = script.downloadUrlToString(apiBaseUrl + '/api/tags') if (data === '') { return; @@ -40,7 +40,7 @@ Script { const modelNames = dataJson.models.map(model => model.model); if (modelNames.length > 0) { - this.models = modelNames.join(','); + models = modelNames.join(','); } } @@ -49,7 +49,7 @@ Script { * It returns a list of objects with config parameters for new OpenAI backends */ function openAiBackendsHook() { - const baseUrl = this.apiBaseUrl + '/v1/chat/completions'; + const baseUrl = apiBaseUrl + '/v1/chat/completions'; return [ {