docs/layouts/shortcodes/generated/openai_configuration.html
| Key | Default | Type | Description |
|---|---|---|---|
| (none) | String | OpenAI API key for authentication. | |
| truncated-tail |
Enum
| Action to handle context overflows.
Possible values:
| |
| (none) | Long | The size of the embedding result array. | |
| (none) | String | Full URL of the OpenAI API endpoint, e.g., https://api.openai.com/v1/chat/completions or https://api.openai.com/v1/embeddings |
|
| RETRY |
Enum
| Strategy for handling errors during model requests.
Possible values:
| |
| (none) | Integer | Max number of tokens for context. context-overflow-action would be triggered if this threshold is exceeded. | |
| (none) | Long | The maximum number of tokens that can be generated in the chat completion. | |
| (none) | String | Model name, e.g., gpt-3.5-turbo, text-embedding-ada-002. |
|
| (none) | Long | How many chat completion choices to generate for each input message. Note that you will be charged based on the number of generated tokens across all of the choices. Keep n as 1 to minimize costs. | |
| (none) | Double | Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics. | |
| (none) |
Enum
| The format of the response, e.g., 'text' or 'json_object'.
Possible values:
| |
| FAILOVER |
Enum
| Fallback strategy to employ if the retry attempts are exhausted. This strategy is applied when error-handling-strategy is set to retry.
Possible values:
| |
| 100 | Integer | Number of retry for OpenAI client requests. | |
| (none) | Long | If specified, the model platform will make a best effort to sample deterministically, such that repeated requests with the same seed and parameters should return the same result. Determinism is not guaranteed. | |
| (none) | String | A CSV list of strings to pass as stop sequences to the model. | |
| "You are a helpful assistant." | String | The system message of a chat. | |
| (none) | Double | Controls the randomness or “creativity” of the output. Typical values are between 0.0 and 1.0. | |
| (none) | Double | The probability cutoff for token selection. Usually, either temperature or topP are specified, but not both. |