Skip to main content
Version: dev

SiliconFlow

Configure DB-GPT to use SiliconFlow's hosted model API. SiliconFlow provides access to multiple open-source models through a unified API, hosted in China.

Prerequisites​

Install dependencies​

uv sync --all-packages \
--extra "base" \
--extra "proxy_openai" \
--extra "rag" \
--extra "storage_chromadb" \
--extra "dbgpts"

Configuration​

Edit configs/dbgpt-proxy-siliconflow.toml:

[models]
[[models.llms]]
name = "Qwen/Qwen2.5-Coder-32B-Instruct"
provider = "proxy/siliconflow"
api_key = "${env:SILICONFLOW_API_KEY}"

[[models.embeddings]]
name = "BAAI/bge-large-zh-v1.5"
provider = "proxy/openai"
api_url = "https://api.siliconflow.cn/v1/embeddings"
api_key = "${env:SILICONFLOW_API_KEY}"

[[models.rerankers]]
name = "BAAI/bge-reranker-v2-m3"
provider = "proxy/siliconflow"
api_key = "${env:SILICONFLOW_API_KEY}"
tip

Set the environment variable:

export SILICONFLOW_API_KEY="your-siliconflow-api-key"

Available models​

SiliconFlow hosts a wide range of open-source models. Some popular choices:

ModelConfig nameNotes
Qwen2.5-Coder-32BQwen/Qwen2.5-Coder-32B-InstructCode-focused
Qwen2.5-72BQwen/Qwen2.5-72B-InstructGeneral purpose
DeepSeek-V3deepseek-ai/DeepSeek-V3Strong reasoning
GLM-4-9BTHUDM/glm-4-9b-chatChinese & English
info

Check SiliconFlow's model list for the latest available models and pricing.

Features​

SiliconFlow configuration also supports rerankers for enhanced RAG retrieval:

[[models.rerankers]]
name = "BAAI/bge-reranker-v2-m3"
provider = "proxy/siliconflow"
api_key = "${env:SILICONFLOW_API_KEY}"

Start the server​

uv run dbgpt start webserver --config configs/dbgpt-proxy-siliconflow.toml

Troubleshooting​

IssueSolution
Authentication failedVerify your SiliconFlow API key
Model not availableCheck SiliconFlow's current model offerings
Slow responsesSome larger models may have higher latency

What's next​