Files
openai.openai-go/batch.go
2026-03-25 22:08:50 +00:00

465 lines
19 KiB
Go

// File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
package openai
import (
"context"
"errors"
"fmt"
"net/http"
"net/url"
"slices"
"github.com/openai/openai-go/v3/internal/apijson"
"github.com/openai/openai-go/v3/internal/apiquery"
"github.com/openai/openai-go/v3/internal/requestconfig"
"github.com/openai/openai-go/v3/option"
"github.com/openai/openai-go/v3/packages/pagination"
"github.com/openai/openai-go/v3/packages/param"
"github.com/openai/openai-go/v3/packages/respjson"
"github.com/openai/openai-go/v3/shared"
"github.com/openai/openai-go/v3/shared/constant"
)
// Create large batches of API requests to run asynchronously.
//
// BatchService contains methods and other services that help with interacting with
// the openai API.
//
// Note, unlike clients, this service does not read variables from the environment
// automatically. You should not instantiate this service directly, and instead use
// the [NewBatchService] method instead.
type BatchService struct {
Options []option.RequestOption
}
// NewBatchService generates a new service that applies the given options to each
// request. These options are applied after the parent client's options (if there
// is one), and before any request-specific options.
func NewBatchService(opts ...option.RequestOption) (r BatchService) {
r = BatchService{}
r.Options = opts
return
}
// Creates and executes a batch from an uploaded file of requests
func (r *BatchService) New(ctx context.Context, body BatchNewParams, opts ...option.RequestOption) (res *Batch, err error) {
opts = slices.Concat(r.Options, opts)
path := "batches"
err = requestconfig.ExecuteNewRequest(ctx, http.MethodPost, path, body, &res, opts...)
return res, err
}
// Retrieves a batch.
func (r *BatchService) Get(ctx context.Context, batchID string, opts ...option.RequestOption) (res *Batch, err error) {
opts = slices.Concat(r.Options, opts)
if batchID == "" {
err = errors.New("missing required batch_id parameter")
return nil, err
}
path := fmt.Sprintf("batches/%s", batchID)
err = requestconfig.ExecuteNewRequest(ctx, http.MethodGet, path, nil, &res, opts...)
return res, err
}
// List your organization's batches.
func (r *BatchService) List(ctx context.Context, query BatchListParams, opts ...option.RequestOption) (res *pagination.CursorPage[Batch], err error) {
var raw *http.Response
opts = slices.Concat(r.Options, opts)
opts = append([]option.RequestOption{option.WithResponseInto(&raw)}, opts...)
path := "batches"
cfg, err := requestconfig.NewRequestConfig(ctx, http.MethodGet, path, query, &res, opts...)
if err != nil {
return nil, err
}
err = cfg.Execute()
if err != nil {
return nil, err
}
res.SetPageConfig(cfg, raw)
return res, nil
}
// List your organization's batches.
func (r *BatchService) ListAutoPaging(ctx context.Context, query BatchListParams, opts ...option.RequestOption) *pagination.CursorPageAutoPager[Batch] {
return pagination.NewCursorPageAutoPager(r.List(ctx, query, opts...))
}
// Cancels an in-progress batch. The batch will be in status `cancelling` for up to
// 10 minutes, before changing to `cancelled`, where it will have partial results
// (if any) available in the output file.
func (r *BatchService) Cancel(ctx context.Context, batchID string, opts ...option.RequestOption) (res *Batch, err error) {
opts = slices.Concat(r.Options, opts)
if batchID == "" {
err = errors.New("missing required batch_id parameter")
return nil, err
}
path := fmt.Sprintf("batches/%s/cancel", batchID)
err = requestconfig.ExecuteNewRequest(ctx, http.MethodPost, path, nil, &res, opts...)
return res, err
}
type Batch struct {
ID string `json:"id" api:"required"`
// The time frame within which the batch should be processed.
CompletionWindow string `json:"completion_window" api:"required"`
// The Unix timestamp (in seconds) for when the batch was created.
CreatedAt int64 `json:"created_at" api:"required"`
// The OpenAI API endpoint used by the batch.
Endpoint string `json:"endpoint" api:"required"`
// The ID of the input file for the batch.
InputFileID string `json:"input_file_id" api:"required"`
// The object type, which is always `batch`.
Object constant.Batch `json:"object" default:"batch"`
// The current status of the batch.
//
// Any of "validating", "failed", "in_progress", "finalizing", "completed",
// "expired", "cancelling", "cancelled".
Status BatchStatus `json:"status" api:"required"`
// The Unix timestamp (in seconds) for when the batch was cancelled.
CancelledAt int64 `json:"cancelled_at"`
// The Unix timestamp (in seconds) for when the batch started cancelling.
CancellingAt int64 `json:"cancelling_at"`
// The Unix timestamp (in seconds) for when the batch was completed.
CompletedAt int64 `json:"completed_at"`
// The ID of the file containing the outputs of requests with errors.
ErrorFileID string `json:"error_file_id"`
Errors BatchErrors `json:"errors"`
// The Unix timestamp (in seconds) for when the batch expired.
ExpiredAt int64 `json:"expired_at"`
// The Unix timestamp (in seconds) for when the batch will expire.
ExpiresAt int64 `json:"expires_at"`
// The Unix timestamp (in seconds) for when the batch failed.
FailedAt int64 `json:"failed_at"`
// The Unix timestamp (in seconds) for when the batch started finalizing.
FinalizingAt int64 `json:"finalizing_at"`
// The Unix timestamp (in seconds) for when the batch started processing.
InProgressAt int64 `json:"in_progress_at"`
// Set of 16 key-value pairs that can be attached to an object. This can be useful
// for storing additional information about the object in a structured format, and
// querying for objects via API or the dashboard.
//
// Keys are strings with a maximum length of 64 characters. Values are strings with
// a maximum length of 512 characters.
Metadata shared.Metadata `json:"metadata" api:"nullable"`
// Model ID used to process the batch, like `gpt-5-2025-08-07`. OpenAI offers a
// wide range of models with different capabilities, performance characteristics,
// and price points. Refer to the
// [model guide](https://platform.openai.com/docs/models) to browse and compare
// available models.
Model string `json:"model"`
// The ID of the file containing the outputs of successfully executed requests.
OutputFileID string `json:"output_file_id"`
// The request counts for different statuses within the batch.
RequestCounts BatchRequestCounts `json:"request_counts"`
// Represents token usage details including input tokens, output tokens, a
// breakdown of output tokens, and the total tokens used. Only populated on batches
// created after September 7, 2025.
Usage BatchUsage `json:"usage"`
// JSON contains metadata for fields, check presence with [respjson.Field.Valid].
JSON struct {
ID respjson.Field
CompletionWindow respjson.Field
CreatedAt respjson.Field
Endpoint respjson.Field
InputFileID respjson.Field
Object respjson.Field
Status respjson.Field
CancelledAt respjson.Field
CancellingAt respjson.Field
CompletedAt respjson.Field
ErrorFileID respjson.Field
Errors respjson.Field
ExpiredAt respjson.Field
ExpiresAt respjson.Field
FailedAt respjson.Field
FinalizingAt respjson.Field
InProgressAt respjson.Field
Metadata respjson.Field
Model respjson.Field
OutputFileID respjson.Field
RequestCounts respjson.Field
Usage respjson.Field
ExtraFields map[string]respjson.Field
raw string
} `json:"-"`
}
// Returns the unmodified JSON received from the API
func (r Batch) RawJSON() string { return r.JSON.raw }
func (r *Batch) UnmarshalJSON(data []byte) error {
return apijson.UnmarshalRoot(data, r)
}
// The current status of the batch.
type BatchStatus string
const (
BatchStatusValidating BatchStatus = "validating"
BatchStatusFailed BatchStatus = "failed"
BatchStatusInProgress BatchStatus = "in_progress"
BatchStatusFinalizing BatchStatus = "finalizing"
BatchStatusCompleted BatchStatus = "completed"
BatchStatusExpired BatchStatus = "expired"
BatchStatusCancelling BatchStatus = "cancelling"
BatchStatusCancelled BatchStatus = "cancelled"
)
type BatchErrors struct {
Data []BatchError `json:"data"`
// The object type, which is always `list`.
Object string `json:"object"`
// JSON contains metadata for fields, check presence with [respjson.Field.Valid].
JSON struct {
Data respjson.Field
Object respjson.Field
ExtraFields map[string]respjson.Field
raw string
} `json:"-"`
}
// Returns the unmodified JSON received from the API
func (r BatchErrors) RawJSON() string { return r.JSON.raw }
func (r *BatchErrors) UnmarshalJSON(data []byte) error {
return apijson.UnmarshalRoot(data, r)
}
type BatchError struct {
// An error code identifying the error type.
Code string `json:"code"`
// The line number of the input file where the error occurred, if applicable.
Line int64 `json:"line" api:"nullable"`
// A human-readable message providing more details about the error.
Message string `json:"message"`
// The name of the parameter that caused the error, if applicable.
Param string `json:"param" api:"nullable"`
// JSON contains metadata for fields, check presence with [respjson.Field.Valid].
JSON struct {
Code respjson.Field
Line respjson.Field
Message respjson.Field
Param respjson.Field
ExtraFields map[string]respjson.Field
raw string
} `json:"-"`
}
// Returns the unmodified JSON received from the API
func (r BatchError) RawJSON() string { return r.JSON.raw }
func (r *BatchError) UnmarshalJSON(data []byte) error {
return apijson.UnmarshalRoot(data, r)
}
// The request counts for different statuses within the batch.
type BatchRequestCounts struct {
// Number of requests that have been completed successfully.
Completed int64 `json:"completed" api:"required"`
// Number of requests that have failed.
Failed int64 `json:"failed" api:"required"`
// Total number of requests in the batch.
Total int64 `json:"total" api:"required"`
// JSON contains metadata for fields, check presence with [respjson.Field.Valid].
JSON struct {
Completed respjson.Field
Failed respjson.Field
Total respjson.Field
ExtraFields map[string]respjson.Field
raw string
} `json:"-"`
}
// Returns the unmodified JSON received from the API
func (r BatchRequestCounts) RawJSON() string { return r.JSON.raw }
func (r *BatchRequestCounts) UnmarshalJSON(data []byte) error {
return apijson.UnmarshalRoot(data, r)
}
// Represents token usage details including input tokens, output tokens, a
// breakdown of output tokens, and the total tokens used. Only populated on batches
// created after September 7, 2025.
type BatchUsage struct {
// The number of input tokens.
InputTokens int64 `json:"input_tokens" api:"required"`
// A detailed breakdown of the input tokens.
InputTokensDetails BatchUsageInputTokensDetails `json:"input_tokens_details" api:"required"`
// The number of output tokens.
OutputTokens int64 `json:"output_tokens" api:"required"`
// A detailed breakdown of the output tokens.
OutputTokensDetails BatchUsageOutputTokensDetails `json:"output_tokens_details" api:"required"`
// The total number of tokens used.
TotalTokens int64 `json:"total_tokens" api:"required"`
// JSON contains metadata for fields, check presence with [respjson.Field.Valid].
JSON struct {
InputTokens respjson.Field
InputTokensDetails respjson.Field
OutputTokens respjson.Field
OutputTokensDetails respjson.Field
TotalTokens respjson.Field
ExtraFields map[string]respjson.Field
raw string
} `json:"-"`
}
// Returns the unmodified JSON received from the API
func (r BatchUsage) RawJSON() string { return r.JSON.raw }
func (r *BatchUsage) UnmarshalJSON(data []byte) error {
return apijson.UnmarshalRoot(data, r)
}
// A detailed breakdown of the input tokens.
type BatchUsageInputTokensDetails struct {
// The number of tokens that were retrieved from the cache.
// [More on prompt caching](https://platform.openai.com/docs/guides/prompt-caching).
CachedTokens int64 `json:"cached_tokens" api:"required"`
// JSON contains metadata for fields, check presence with [respjson.Field.Valid].
JSON struct {
CachedTokens respjson.Field
ExtraFields map[string]respjson.Field
raw string
} `json:"-"`
}
// Returns the unmodified JSON received from the API
func (r BatchUsageInputTokensDetails) RawJSON() string { return r.JSON.raw }
func (r *BatchUsageInputTokensDetails) UnmarshalJSON(data []byte) error {
return apijson.UnmarshalRoot(data, r)
}
// A detailed breakdown of the output tokens.
type BatchUsageOutputTokensDetails struct {
// The number of reasoning tokens.
ReasoningTokens int64 `json:"reasoning_tokens" api:"required"`
// JSON contains metadata for fields, check presence with [respjson.Field.Valid].
JSON struct {
ReasoningTokens respjson.Field
ExtraFields map[string]respjson.Field
raw string
} `json:"-"`
}
// Returns the unmodified JSON received from the API
func (r BatchUsageOutputTokensDetails) RawJSON() string { return r.JSON.raw }
func (r *BatchUsageOutputTokensDetails) UnmarshalJSON(data []byte) error {
return apijson.UnmarshalRoot(data, r)
}
type BatchNewParams struct {
// The time frame within which the batch should be processed. Currently only `24h`
// is supported.
//
// Any of "24h".
CompletionWindow BatchNewParamsCompletionWindow `json:"completion_window,omitzero" api:"required"`
// The endpoint to be used for all requests in the batch. Currently
// `/v1/responses`, `/v1/chat/completions`, `/v1/embeddings`, `/v1/completions`,
// `/v1/moderations`, `/v1/images/generations`, `/v1/images/edits`, and
// `/v1/videos` are supported. Note that `/v1/embeddings` batches are also
// restricted to a maximum of 50,000 embedding inputs across all requests in the
// batch.
//
// Any of "/v1/responses", "/v1/chat/completions", "/v1/embeddings",
// "/v1/completions", "/v1/moderations", "/v1/images/generations",
// "/v1/images/edits", "/v1/videos".
Endpoint BatchNewParamsEndpoint `json:"endpoint,omitzero" api:"required"`
// The ID of an uploaded file that contains requests for the new batch.
//
// See [upload file](https://platform.openai.com/docs/api-reference/files/create)
// for how to upload a file.
//
// Your input file must be formatted as a
// [JSONL file](https://platform.openai.com/docs/api-reference/batch/request-input),
// and must be uploaded with the purpose `batch`. The file can contain up to 50,000
// requests, and can be up to 200 MB in size.
InputFileID string `json:"input_file_id" api:"required"`
// Set of 16 key-value pairs that can be attached to an object. This can be useful
// for storing additional information about the object in a structured format, and
// querying for objects via API or the dashboard.
//
// Keys are strings with a maximum length of 64 characters. Values are strings with
// a maximum length of 512 characters.
Metadata shared.Metadata `json:"metadata,omitzero"`
// The expiration policy for the output and/or error file that are generated for a
// batch.
OutputExpiresAfter BatchNewParamsOutputExpiresAfter `json:"output_expires_after,omitzero"`
paramObj
}
func (r BatchNewParams) MarshalJSON() (data []byte, err error) {
type shadow BatchNewParams
return param.MarshalObject(r, (*shadow)(&r))
}
func (r *BatchNewParams) UnmarshalJSON(data []byte) error {
return apijson.UnmarshalRoot(data, r)
}
// The time frame within which the batch should be processed. Currently only `24h`
// is supported.
type BatchNewParamsCompletionWindow string
const (
BatchNewParamsCompletionWindow24h BatchNewParamsCompletionWindow = "24h"
)
// The endpoint to be used for all requests in the batch. Currently
// `/v1/responses`, `/v1/chat/completions`, `/v1/embeddings`, `/v1/completions`,
// `/v1/moderations`, `/v1/images/generations`, `/v1/images/edits`, and
// `/v1/videos` are supported. Note that `/v1/embeddings` batches are also
// restricted to a maximum of 50,000 embedding inputs across all requests in the
// batch.
type BatchNewParamsEndpoint string
const (
BatchNewParamsEndpointV1Responses BatchNewParamsEndpoint = "/v1/responses"
BatchNewParamsEndpointV1ChatCompletions BatchNewParamsEndpoint = "/v1/chat/completions"
BatchNewParamsEndpointV1Embeddings BatchNewParamsEndpoint = "/v1/embeddings"
BatchNewParamsEndpointV1Completions BatchNewParamsEndpoint = "/v1/completions"
BatchNewParamsEndpointV1Moderations BatchNewParamsEndpoint = "/v1/moderations"
BatchNewParamsEndpointV1ImagesGenerations BatchNewParamsEndpoint = "/v1/images/generations"
BatchNewParamsEndpointV1ImagesEdits BatchNewParamsEndpoint = "/v1/images/edits"
BatchNewParamsEndpointV1Videos BatchNewParamsEndpoint = "/v1/videos"
)
// The expiration policy for the output and/or error file that are generated for a
// batch.
//
// The properties Anchor, Seconds are required.
type BatchNewParamsOutputExpiresAfter struct {
// The number of seconds after the anchor time that the file will expire. Must be
// between 3600 (1 hour) and 2592000 (30 days).
Seconds int64 `json:"seconds" api:"required"`
// Anchor timestamp after which the expiration policy applies. Supported anchors:
// `created_at`. Note that the anchor is the file creation time, not the time the
// batch is created.
//
// This field can be elided, and will marshal its zero value as "created_at".
Anchor constant.CreatedAt `json:"anchor" default:"created_at"`
paramObj
}
func (r BatchNewParamsOutputExpiresAfter) MarshalJSON() (data []byte, err error) {
type shadow BatchNewParamsOutputExpiresAfter
return param.MarshalObject(r, (*shadow)(&r))
}
func (r *BatchNewParamsOutputExpiresAfter) UnmarshalJSON(data []byte) error {
return apijson.UnmarshalRoot(data, r)
}
type BatchListParams struct {
// A cursor for use in pagination. `after` is an object ID that defines your place
// in the list. For instance, if you make a list request and receive 100 objects,
// ending with obj_foo, your subsequent call can include after=obj_foo in order to
// fetch the next page of the list.
After param.Opt[string] `query:"after,omitzero" json:"-"`
// A limit on the number of objects to be returned. Limit can range between 1 and
// 100, and the default is 20.
Limit param.Opt[int64] `query:"limit,omitzero" json:"-"`
paramObj
}
// URLQuery serializes [BatchListParams]'s query parameters as `url.Values`.
func (r BatchListParams) URLQuery() (v url.Values, err error) {
return apiquery.MarshalWithSettings(r, apiquery.QuerySettings{
ArrayFormat: apiquery.ArrayQueryFormatBrackets,
NestedFormat: apiquery.NestedQueryFormatBrackets,
})
}