AgentCompletionsRequestPayload

AgentCompletionsRequestPayload
AgentCompletionsRequestPayload

Request to generate completions from an agent.

JSON Example
{
    "max_tokens": 64,
    "model": "model-to-use",
    "n": 1,
    "prompt": "hello",
    "stop": [
        "END"
    ],
    "temperature": 0
}
boolean
create_session
Optional

If true, the request creates a new agent session and the LLM interaction is stored as context for subsequent agent interactions when using the generated session.

integer
seed
Optional

Seed to propagate to the LLM for making repeated requests with the same seed as deterministic as possible. Note that this feature is in beta for most inference servers.

number
temperature
Optional
Constraints: minimum: 0 maximum: 2 default: 0

temperature

integer
n
Optional
Constraints: minimum: 1 default: 1

n

array of string
stop
Optional
Constraints: minItems: 1

stop

integer
max_tokens
Optional
Constraints: minimum: 1

max_tokens

boolean
stream
Optional

stream

string
model
Optional

Optional ID of the model to use. If provided, it must match the model specified in the agent configuration. Unless the client needs to validate that the specified model is in use by the agent, do not specify this value and the API will choose the correct model. For compatibility with the OpenAI client SDK, this parameter may either be unset or an empty string may be used to indicate the use of the agent default configuration.

string
prompt
Required

prompt

string As uuid As uuid
store_in_session
Optional

If set, use and extend the context stored in the given session for all LLM interactions.