Skip to main content
Using Deno in production at your company? Earn free Deno merch.
Give us feedback
Module

x/openai/resources/beta/threads/mod.ts>Run

Deno build of the official Typescript library for the OpenAI API.
Extremely Popular
Go to Latest
namespace Run
import { Run } from "https://deno.land/x/openai@v4.38.5/resources/beta/threads/mod.ts";

Interfaces

Details on why the run is incomplete. Will be null if the run is not incomplete.

The last error associated with this run. Will be null if there are no errors.

Details on the action required to continue the run. Will be null if no action is required.

Details on the tool outputs needed for this run to continue.

Controls for how a thread will be truncated prior to the run. Use this to control the intial context window of the run.

Usage statistics related to the run. This value will be null if the run is not in a terminal state (i.e. in_progress, queued, etc.).

interface Run
import { type Run } from "https://deno.land/x/openai@v4.38.5/resources/beta/threads/mod.ts";

Represents an execution run on a thread.

Properties

id: string

The identifier, which can be referenced in API endpoints.

assistant_id: string

The ID of the assistant used for execution of this run.

cancelled_at: number | null

The Unix timestamp (in seconds) for when the run was cancelled.

completed_at: number | null

The Unix timestamp (in seconds) for when the run was completed.

created_at: number

The Unix timestamp (in seconds) for when the run was created.

expires_at: number | null

The Unix timestamp (in seconds) for when the run will expire.

failed_at: number | null

The Unix timestamp (in seconds) for when the run failed.

incomplete_details: Run.IncompleteDetails | null

Details on why the run is incomplete. Will be null if the run is not incomplete.

instructions: string

The instructions that the assistant used for this run.

last_error: Run.LastError | null

The last error associated with this run. Will be null if there are no errors.

max_completion_tokens: number | null

The maximum number of completion tokens specified to have been used over the course of the run.

max_prompt_tokens: number | null

The maximum number of prompt tokens specified to have been used over the course of the run.

metadata: unknown | null

Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format. Keys can be a maximum of 64 characters long and values can be a maxium of 512 characters long.

model: string

The model that the assistant used for this run.

object: "thread.run"

The object type, which is always thread.run.

required_action: Run.RequiredAction | null

Details on the action required to continue the run. Will be null if no action is required.

response_format: ThreadsAPI.AssistantResponseFormatOption | null

Specifies the format that the model must output. Compatible with GPT-4 Turbo and all GPT-3.5 Turbo models since gpt-3.5-turbo-1106.

Setting to { "type": "json_object" } enables JSON mode, which guarantees the message the model generates is valid JSON.

Important: when using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.

started_at: number | null

The Unix timestamp (in seconds) for when the run was started.

status: RunStatus

The status of the run, which can be either queued, in_progress, requires_action, cancelling, cancelled, failed, completed, or expired.

thread_id: string

The ID of the thread that was executed on as a part of this run.

tool_choice: ThreadsAPI.AssistantToolChoiceOption | null

Controls which (if any) tool is called by the model. none means the model will not call any tools and instead generates a message. auto is the default value and means the model can pick between generating a message or calling a tool. Specifying a particular tool like {"type": "file_search"} or {"type": "function", "function": {"name": "my_function"}} forces the model to call that tool.

tools: Array<AssistantsAPI.AssistantTool>

The list of tools that the assistant used for this run.

truncation_strategy: Run.TruncationStrategy | null

Controls for how a thread will be truncated prior to the run. Use this to control the intial context window of the run.

usage: Run.Usage | null

Usage statistics related to the run. This value will be null if the run is not in a terminal state (i.e. in_progress, queued, etc.).

optional
temperature: number | null

The sampling temperature used for this run. If not set, defaults to 1.

optional
top_p: number | null

The nucleus sampling value used for this run. If not set, defaults to 1.