| Name | Type | Description | Notes |
|---|---|---|---|
| messages | List<ChatCompletionSystemMessageParam> | A list of the previous chat messages for context. | |
| model | String | The language model to chat with. If you are optimizing for speed, try mistralai/Mistral-7B-Instruct-v0.1. For quality, try NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO | [optional] |
| stream | Boolean | Whether or not to stream data-only server-sent events as they become available. | [optional] |
| maxTokens | Integer | Maximum number of completion tokens the model should generate. | [optional] |
| temperature | BigDecimal | Adjusts the "creativity" of the model. Lower values make the model more deterministic and repetitive, while higher values make the model more random and creative. | [optional] |
| minP | BigDecimal | This is an alternative to `temperature` that many prefer. Must be in [0, 1]. | [optional] |
| n | BigDecimal | This will return multiple choices for you instead of a single chat completion. | [optional] |
| tools | List<ChatCompletionRequestToolsInner> | The `retrieval` tool type is unique to Telnyx. You may pass a list of embedded storage buckets for retrieval-augmented generation. | [optional] |
| toolChoice | ToolChoiceEnum | [optional] | |
| useBeamSearch | Boolean | Setting this to `true` will allow the model to explore more completion options. This is not supported by OpenAI. | [optional] |
| bestOf | Integer | This is used with `use_beam_search` to determine how many candidate beams to explore. | [optional] |
| lengthPenalty | BigDecimal | This is used with `use_beam_search` to prefer shorter or longer completions. | [optional] |
| earlyStopping | Boolean | This is used with `use_beam_search`. If `true`, generation stops as soon as there are `best_of` complete candidates; if `false`, a heuristic is applied and the generation stops when is it very unlikely to find better candidates. | [optional] |
| frequencyPenalty | BigDecimal | Higher values will penalize the model from repeating the same output tokens. | [optional] |
| presencePenalty | BigDecimal | Higher values will penalize the model from repeating the same output tokens. | [optional] |
| topP | BigDecimal | An alternative to `temperature`. | [optional] |
| openaiApiKey | String | If you are using OpenAI models using our API, this is how you pass along your OpenAI API key. | [optional] |
| Name | Value |
|---|---|
| NONE | "none" |
| AUTO | "auto" |