Sends input to an OpenAI compatible Responses model and returns a unified response object. Supports text and chat-style input, streaming, tools, and (for supported reasoning models) “thinking” via the reasoning object. To send a request to a closed source provider, prefix your model with their provider name, e.g. openai/gpt-5.1.
Token for authentication
The ID of the model to run (e.g., anthropic/claude-opus-4-5, openai/gpt-5.1)
The input to the model. Can be a string (simple prompt) or an array of chat-style messages. For richer multimodal inputs, use the array form with multi-part content.
Maximum number of tokens to generate (counts reasoning + visible output for reasoning models)
Sampling temperature
Whether to stream SSE events
Optional "thinking" controls for supported reasoning models.
Optional tool definitions for function/tool calling
Tool selection behavior
auto, none, required Arbitrary key/value metadata to attach to the request
End-user identifier (if supported)
Optional list of extra fields to include in the response (e.g. logprobs). Example values may include message.output_text.logprobs (provider/model dependent).
Number of most likely tokens to return at each position (if logprobs are included)
Successful response
Unique ID for this response
Type of returned object (usually response)
Unix timestamp of response creation
Model used to generate the response
Output items (messages, reasoning summaries, tool calls, etc.)
Convenience field containing concatenated output text (when applicable)
Token usage details (shape may vary by provider)