id | str | Required | The id/name of the model to use |
name | Optional[str] | None | The display name of the model |
provider | Optional[str] | None | The provider of the model |
frequency_penalty | Optional[float] | None | Penalizes new tokens based on their frequency in the text so far |
presence_penalty | Optional[float] | None | Penalizes new tokens based on whether they appear in the text so far |
response_format | Optional[str] | None | The format of the response |
seed | Optional[int] | None | Random seed for deterministic sampling |
stop | Optional[Union[str, List[str]]] | None | Up to 4 sequences where the API will stop generating further tokens |
stream | bool | True | Whether to stream the response |
temperature | Optional[float] | None | Controls randomness in the model’s output |
top_p | Optional[float] | None | Controls diversity via nucleus sampling |
max_tokens | Optional[int] | None | Maximum number of tokens to generate |
request_params | Optional[Dict[str, Any]] | None | Additional parameters to include in the request |