-
Notifications
You must be signed in to change notification settings - Fork 2.5k
3. Model configuration
Coze Studio is an AI app development platform based on LLM. Before running the Coze Studio open-source version for the first time, you need to clone the project to your local machine and configure the required models. During normal project operations, you can also add new model services as needed or delete unnecessary model services at any time.
The model services supported by Coze Studio are as follows:
- Volcengine Ark | BytePlus ModelArk
- OpenAI
- DeepSeek
- Claude
- Ollama
- Qwen
- Gemini
Before filling out the model configuration file, ensure you have understood the following Important notes:
- Before deleting the model, ensure that it is no longer receiving online traffic.
- Agents or workflows call models based on model IDs. For models that have already been launched, do not modify their IDs; otherwise, it may result in model call failures.
Coze Studio is an AI app development platform based on LLMs. Before deploying and launching the open-source version of Coze Studio for the first time, you need to configure the model service in the Coze Studio project, otherwise, you won't be able to properly select a model during the creation of agents or workflows.
Configure the model at http://localhost:8888/admin/#model-management by adding a new model.
| Platform | Protocol | Base_url | Special Instructions |
|---|---|---|---|
| Volcengine Ark | ark | Volcengine Engine: https://ark.cn-beijing.volces.com/api/v3/ Overseas BytePlus: https://ark.ap-southeast.bytepluses.com/api/v3/ |
None |
| Alibaba Bai Lian | openai or qwen |
https://dashscope.aliyuncs.com/compatible-mode/v1 | The qwen3 series does not support thinking in non-streaming calls. If used, you need to set enable_thinking: false in conn_config. Coze Studio will adapt this capability in future versions. |
| Silicon-based Flow | openai | https://api.siliconflow.cn/v1 | None |
| Other Third-Party API Relay | openai | The address provided in the API document Note that the path usually has a /v1 suffix and does not have a /chat/completions suffix |
If the platform only relays or proxies model services and the model is not an openai model, please configure the protocol according to the documentation in the [Official Model Service] section. |
| Framework | Protocol | Base_url | Special Instructions |
|---|---|---|---|
| Ollama | ollama | http://${ip}:11434 | 1. When the mirror network mode is bridge, localhost in the coze-server mirror is not the localhost of the host. It needs to be modified to the ip of the Ollama deployment machine or http://host.docker.internal:11434. 2. Check the api_key: When the api_key is not set, this parameter is left blank. 3. Confirm whether the firewall of the Ollama-deployed host has opened port 11434. 4. Confirm that the Ollama network has enabled External Exposure. |
| vllm | openai | http://${ip}:8000/v1 (specified when starting the port) | None |
| xinference | openai | http://${ip}:9997/v1 (specified when starting the port) | None |
| sglang | openai | http://${ip}:35140/v1 (specified when starting the port) | None |
| LMStudio | openai | http://${ip}:${port}/v1 | None |
| Model | Protocol | Base_url | Special Instructions |
|---|---|---|---|
| Doubao | ark | https://ark.cn-beijing.volces.com/api/v3/ | None |
| OpenAI | openai | https://api.openai.com/v1 | Check the by_azure field configuration. If the model is a model service provided by Microsoft Azure, this parameter should be set to true. |
| Deepseek | deepseek | https://api.deepseek.com/ | None |
| Qwen | qwen | https://dashscope.aliyuncs.com/compatible-mode/v1 | The qwen3 series does not support thinking in non-streaming calls. If used, you need to set enable_thinking: false in conn_config. Coze Studio will adapt this capability in future versions. |
| Gemini | gemini | https://generativelanguage.googleapis.com/ | None |
| Claude | claude | https://api.anthropic.com/v1/ | None |