跳到主要内容
版本:dev

Baidu Wenxin Proxy LLM Configuration

Baidu Wenxin proxy LLM configuration.

Details can be found in:
https://cloud.baidu.com/doc/WENXINWORKSHOP/s/clntwmv7t

Parameters

NameTypeRequiredDescription
namestring
The name of the model.
backendstring
The real model name to pass to the provider, default is None. If backend is None, use name as the real model name.
providerstring
The provider of the model. If model is deployed in local, this is the inference type. If model is deployed in third-party service, this is platform name('proxy/<platform>')
Defaults:proxy/wenxin
verboseboolean
Show verbose output.
Defaults:False
concurrencyinteger
Model concurrency limit
Defaults:100
prompt_templatestring
Prompt template. If None, the prompt template is automatically determined from model. Just for local deployment.
context_lengthinteger
The context length of the OpenAI API. If None, it is determined by the model.
api_keystring
The API key of the Wenxin API.
Defaults:${env:WEN_XIN_API_KEY}
api_secretstring
The API secret key of the Wenxin API.
Defaults:${env:WEN_XIN_API_SECRET}