AI/ML API Proxy LLM Configuration
OpenAI-compatible chat completion request schema.
Details can be found in:
https://api.aimlapi.com/docs-public
Parameters
| Name | Type | Required | Description |
|---|---|---|---|
model | string | â | ID of the model to use. |
messages | array | â | List of messages comprising the conversation. |
max_completion_tokens | integer | â | Maximum number of tokens to generate for completion. |
max_tokens | integer | â | Alias for max_completion_tokens. |
stream | boolean | â | Whether to stream back partial progress. |
stream_options | object | â | Additional options to control streaming behavior. |
tools | array | â | List of tools (functions or APIs) the model may call. |
tool_choice | object | â | Which tool the model should call, if any. |
parallel_tool_calls | boolean | â | Whether tools can be called in parallel. |
n | integer | â | How many completions to generate for each prompt. |
stop | array|string | â | Sequences where the model will stop generating further tokens. |
logprobs | boolean | â | Whether to include log probabilities for tokens. |
top_logprobs | integer | â | Number of most likely tokens to return logprobs for. |
logit_bias | object | â | Modify likelihood of specified tokens appearing in the completion. |
frequency_penalty | number | â | How much to penalize new tokens based on frequency. |
presence_penalty | number | â | How much to penalize new tokens based on whether they appear in the text so far. |
seed | integer | â | Seed for sampling (for reproducibility). |
temperature | number | â | Sampling temperature to use (higher = more random). |
top_p | number | â | Nucleus sampling (top-p) cutoff value. |
response_format | object|string | â | Format to return the completion in, such as 'json' or 'text'. |