diff --git a/.stats.yml b/.stats.yml
index 39413df445..f368bc881d 100644
--- a/.stats.yml
+++ b/.stats.yml
@@ -1,2 +1,2 @@
configured_endpoints: 68
-openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai-7b0a5d715d94f75ac7795bd4d2175a0e3243af9b935a86c273f371e45583140f.yml
+openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai-2f8ca92b9b1879fd535b685e4767338413fcd533d42f3baac13a9c41da3fce35.yml
diff --git a/api.md b/api.md
index 4600adf77a..3e88faab91 100644
--- a/api.md
+++ b/api.md
@@ -54,6 +54,7 @@ from openai.types.chat import (
ChatCompletionMessageToolCall,
ChatCompletionModality,
ChatCompletionNamedToolChoice,
+ ChatCompletionPredictionContent,
ChatCompletionRole,
ChatCompletionStreamOptions,
ChatCompletionSystemMessageParam,
@@ -93,7 +94,7 @@ Methods:
- client.files.create(\*\*params) -> FileObject
- client.files.retrieve(file_id) -> FileObject
-- client.files.list(\*\*params) -> SyncPage[FileObject]
+- client.files.list(\*\*params) -> SyncCursorPage[FileObject]
- client.files.delete(file_id) -> FileDeleted
- client.files.content(file_id) -> HttpxBinaryResponseContent
- client.files.retrieve_content(file_id) -> str
diff --git a/src/openai/resources/audio/speech.py b/src/openai/resources/audio/speech.py
index 6085ae8afe..09faaddda6 100644
--- a/src/openai/resources/audio/speech.py
+++ b/src/openai/resources/audio/speech.py
@@ -70,13 +70,13 @@ def create(
input: The text to generate audio for. The maximum length is 4096 characters.
model:
- One of the available [TTS models](https://platform.openai.com/docs/models/tts):
+ One of the available [TTS models](https://platform.openai.com/docs/models#tts):
`tts-1` or `tts-1-hd`
voice: The voice to use when generating the audio. Supported voices are `alloy`,
`echo`, `fable`, `onyx`, `nova`, and `shimmer`. Previews of the voices are
available in the
- [Text to speech guide](https://platform.openai.com/docs/guides/text-to-speech/voice-options).
+ [Text to speech guide](https://platform.openai.com/docs/guides/text-to-speech#voice-options).
response_format: The format to audio in. Supported formats are `mp3`, `opus`, `aac`, `flac`,
`wav`, and `pcm`.
@@ -154,13 +154,13 @@ async def create(
input: The text to generate audio for. The maximum length is 4096 characters.
model:
- One of the available [TTS models](https://platform.openai.com/docs/models/tts):
+ One of the available [TTS models](https://platform.openai.com/docs/models#tts):
`tts-1` or `tts-1-hd`
voice: The voice to use when generating the audio. Supported voices are `alloy`,
`echo`, `fable`, `onyx`, `nova`, and `shimmer`. Previews of the voices are
available in the
- [Text to speech guide](https://platform.openai.com/docs/guides/text-to-speech/voice-options).
+ [Text to speech guide](https://platform.openai.com/docs/guides/text-to-speech#voice-options).
response_format: The format to audio in. Supported formats are `mp3`, `opus`, `aac`, `flac`,
`wav`, and `pcm`.
diff --git a/src/openai/resources/audio/transcriptions.py b/src/openai/resources/audio/transcriptions.py
index ccff507a41..9fad66ed35 100644
--- a/src/openai/resources/audio/transcriptions.py
+++ b/src/openai/resources/audio/transcriptions.py
@@ -82,7 +82,7 @@ def create(
prompt: An optional text to guide the model's style or continue a previous audio
segment. The
- [prompt](https://platform.openai.com/docs/guides/speech-to-text/prompting)
+ [prompt](https://platform.openai.com/docs/guides/speech-to-text#prompting)
should match the audio language.
response_format: The format of the output, in one of these options: `json`, `text`, `srt`,
@@ -194,7 +194,7 @@ async def create(
prompt: An optional text to guide the model's style or continue a previous audio
segment. The
- [prompt](https://platform.openai.com/docs/guides/speech-to-text/prompting)
+ [prompt](https://platform.openai.com/docs/guides/speech-to-text#prompting)
should match the audio language.
response_format: The format of the output, in one of these options: `json`, `text`, `srt`,
diff --git a/src/openai/resources/audio/translations.py b/src/openai/resources/audio/translations.py
index 27475f1a59..feaeea6e09 100644
--- a/src/openai/resources/audio/translations.py
+++ b/src/openai/resources/audio/translations.py
@@ -74,7 +74,7 @@ def create(
prompt: An optional text to guide the model's style or continue a previous audio
segment. The
- [prompt](https://platform.openai.com/docs/guides/speech-to-text/prompting)
+ [prompt](https://platform.openai.com/docs/guides/speech-to-text#prompting)
should be in English.
response_format: The format of the output, in one of these options: `json`, `text`, `srt`,
@@ -171,7 +171,7 @@ async def create(
prompt: An optional text to guide the model's style or continue a previous audio
segment. The
- [prompt](https://platform.openai.com/docs/guides/speech-to-text/prompting)
+ [prompt](https://platform.openai.com/docs/guides/speech-to-text#prompting)
should be in English.
response_format: The format of the output, in one of these options: `json`, `text`, `srt`,
diff --git a/src/openai/resources/beta/assistants.py b/src/openai/resources/beta/assistants.py
index 2ebef183b6..7df212f155 100644
--- a/src/openai/resources/beta/assistants.py
+++ b/src/openai/resources/beta/assistants.py
@@ -79,8 +79,8 @@ def create(
model: ID of the model to use. You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
description: The description of the assistant. The maximum length is 512 characters.
@@ -95,8 +95,8 @@ def create(
name: The name of the assistant. The maximum length is 256 characters.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -239,14 +239,14 @@ def update(
model: ID of the model to use. You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
name: The name of the assistant. The maximum length is 256 characters.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -344,8 +344,8 @@ def list(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
limit: A limit on the number of objects to be returned. Limit can range between 1 and
100, and the default is 20.
@@ -465,8 +465,8 @@ async def create(
model: ID of the model to use. You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
description: The description of the assistant. The maximum length is 512 characters.
@@ -481,8 +481,8 @@ async def create(
name: The name of the assistant. The maximum length is 256 characters.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -625,14 +625,14 @@ async def update(
model: ID of the model to use. You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
name: The name of the assistant. The maximum length is 256 characters.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -730,8 +730,8 @@ def list(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
limit: A limit on the number of objects to be returned. Limit can range between 1 and
100, and the default is 20.
diff --git a/src/openai/resources/beta/threads/messages.py b/src/openai/resources/beta/threads/messages.py
index 9e6ae8811a..3c25449664 100644
--- a/src/openai/resources/beta/threads/messages.py
+++ b/src/openai/resources/beta/threads/messages.py
@@ -218,8 +218,8 @@ def list(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
limit: A limit on the number of objects to be returned. Limit can range between 1 and
100, and the default is 20.
@@ -492,8 +492,8 @@ def list(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
limit: A limit on the number of objects to be returned. Limit can range between 1 and
100, and the default is 20.
diff --git a/src/openai/resources/beta/threads/runs/runs.py b/src/openai/resources/beta/threads/runs/runs.py
index 68efaf1782..5c97af0e2e 100644
--- a/src/openai/resources/beta/threads/runs/runs.py
+++ b/src/openai/resources/beta/threads/runs/runs.py
@@ -111,7 +111,7 @@ def create(
to fetch the file search result content.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
additional_instructions: Appends additional instructions at the end of the instructions for the run. This
@@ -147,12 +147,12 @@ def create(
assistant will be used.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -255,7 +255,7 @@ def create(
to fetch the file search result content.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
additional_instructions: Appends additional instructions at the end of the instructions for the run. This
@@ -291,12 +291,12 @@ def create(
assistant will be used.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -395,7 +395,7 @@ def create(
to fetch the file search result content.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
additional_instructions: Appends additional instructions at the end of the instructions for the run. This
@@ -431,12 +431,12 @@ def create(
assistant will be used.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -663,8 +663,8 @@ def list(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
limit: A limit on the number of objects to be returned. Limit can range between 1 and
100, and the default is 20.
@@ -959,7 +959,7 @@ async def create(
to fetch the file search result content.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
additional_instructions: Appends additional instructions at the end of the instructions for the run. This
@@ -995,12 +995,12 @@ async def create(
assistant will be used.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -1103,7 +1103,7 @@ async def create(
to fetch the file search result content.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
additional_instructions: Appends additional instructions at the end of the instructions for the run. This
@@ -1139,12 +1139,12 @@ async def create(
assistant will be used.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -1243,7 +1243,7 @@ async def create(
to fetch the file search result content.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
additional_instructions: Appends additional instructions at the end of the instructions for the run. This
@@ -1279,12 +1279,12 @@ async def create(
assistant will be used.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -1511,8 +1511,8 @@ def list(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
limit: A limit on the number of objects to be returned. Limit can range between 1 and
100, and the default is 20.
diff --git a/src/openai/resources/beta/threads/runs/steps.py b/src/openai/resources/beta/threads/runs/steps.py
index 5d6d55f9d9..9bd91e39e0 100644
--- a/src/openai/resources/beta/threads/runs/steps.py
+++ b/src/openai/resources/beta/threads/runs/steps.py
@@ -68,7 +68,7 @@ def retrieve(
to fetch the file search result content.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
extra_headers: Send extra headers
@@ -126,15 +126,15 @@ def list(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
include: A list of additional fields to include in the response. Currently the only
supported value is `step_details.tool_calls[*].file_search.results[*].content`
to fetch the file search result content.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
limit: A limit on the number of objects to be returned. Limit can range between 1 and
@@ -222,7 +222,7 @@ async def retrieve(
to fetch the file search result content.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
extra_headers: Send extra headers
@@ -280,15 +280,15 @@ def list(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
include: A list of additional fields to include in the response. Currently the only
supported value is `step_details.tool_calls[*].file_search.results[*].content`
to fetch the file search result content.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
limit: A limit on the number of objects to be returned. Limit can range between 1 and
diff --git a/src/openai/resources/beta/threads/threads.py b/src/openai/resources/beta/threads/threads.py
index 17f5c6970b..728d375aa6 100644
--- a/src/openai/resources/beta/threads/threads.py
+++ b/src/openai/resources/beta/threads/threads.py
@@ -317,12 +317,12 @@ def create_and_run(
assistant will be used.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -451,12 +451,12 @@ def create_and_run(
assistant will be used.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -581,12 +581,12 @@ def create_and_run(
assistant will be used.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -971,12 +971,12 @@ async def create_and_run(
assistant will be used.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -1105,12 +1105,12 @@ async def create_and_run(
assistant will be used.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -1235,12 +1235,12 @@ async def create_and_run(
assistant will be used.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
diff --git a/src/openai/resources/beta/vector_stores/file_batches.py b/src/openai/resources/beta/vector_stores/file_batches.py
index 34fcd8c61b..2d4cec3ce8 100644
--- a/src/openai/resources/beta/vector_stores/file_batches.py
+++ b/src/openai/resources/beta/vector_stores/file_batches.py
@@ -201,8 +201,8 @@ def list_files(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
filter: Filter by file status. One of `in_progress`, `completed`, `failed`, `cancelled`.
@@ -422,8 +422,8 @@ def list_files(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
filter: Filter by file status. One of `in_progress`, `completed`, `failed`, `cancelled`.
diff --git a/src/openai/resources/beta/vector_stores/files.py b/src/openai/resources/beta/vector_stores/files.py
index e96b492ac0..d633985e0d 100644
--- a/src/openai/resources/beta/vector_stores/files.py
+++ b/src/openai/resources/beta/vector_stores/files.py
@@ -162,8 +162,8 @@ def list(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
filter: Filter by file status. One of `in_progress`, `completed`, `failed`, `cancelled`.
@@ -384,8 +384,8 @@ def list(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
filter: Filter by file status. One of `in_progress`, `completed`, `failed`, `cancelled`.
diff --git a/src/openai/resources/beta/vector_stores/vector_stores.py b/src/openai/resources/beta/vector_stores/vector_stores.py
index d69add7b26..61a2eadc7b 100644
--- a/src/openai/resources/beta/vector_stores/vector_stores.py
+++ b/src/openai/resources/beta/vector_stores/vector_stores.py
@@ -251,8 +251,8 @@ def list(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
limit: A limit on the number of objects to be returned. Limit can range between 1 and
100, and the default is 20.
@@ -529,8 +529,8 @@ def list(
before: A cursor for use in pagination. `before` is an object ID that defines your place
in the list. For instance, if you make a list request and receive 100 objects,
- ending with obj_foo, your subsequent call can include before=obj_foo in order to
- fetch the previous page of the list.
+ starting with obj_foo, your subsequent call can include before=obj_foo in order
+ to fetch the previous page of the list.
limit: A limit on the number of objects to be returned. Limit can range between 1 and
100, and the default is 20.
diff --git a/src/openai/resources/chat/completions.py b/src/openai/resources/chat/completions.py
index 03919aab2f..84e6cf9b72 100644
--- a/src/openai/resources/chat/completions.py
+++ b/src/openai/resources/chat/completions.py
@@ -31,6 +31,7 @@
from ...types.chat.chat_completion_audio_param import ChatCompletionAudioParam
from ...types.chat.chat_completion_message_param import ChatCompletionMessageParam
from ...types.chat.chat_completion_stream_options_param import ChatCompletionStreamOptionsParam
+from ...types.chat.chat_completion_prediction_content_param import ChatCompletionPredictionContentParam
from ...types.chat.chat_completion_tool_choice_option_param import ChatCompletionToolChoiceOptionParam
__all__ = ["Completions", "AsyncCompletions"]
@@ -74,6 +75,7 @@ def create(
modalities: Optional[List[ChatCompletionModality]] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
+ prediction: Optional[ChatCompletionPredictionContentParam] | NotGiven = NOT_GIVEN,
presence_penalty: Optional[float] | NotGiven = NOT_GIVEN,
response_format: completion_create_params.ResponseFormat | NotGiven = NOT_GIVEN,
seed: Optional[int] | NotGiven = NOT_GIVEN,
@@ -111,7 +113,7 @@ def create(
[audio](https://platform.openai.com/docs/guides/audio).
model: ID of the model to use. See the
- [model endpoint compatibility](https://platform.openai.com/docs/models/model-endpoint-compatibility)
+ [model endpoint compatibility](https://platform.openai.com/docs/models#model-endpoint-compatibility)
table for details on which models work with the Chat API.
audio: Parameters for audio output. Required when audio output is requested with
@@ -122,7 +124,7 @@ def create(
existing frequency in the text so far, decreasing the model's likelihood to
repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
function_call: Deprecated in favor of `tool_choice`.
@@ -183,19 +185,22 @@ def create(
choices. Keep `n` as `1` to minimize costs.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
+ prediction: Static predicted output content, such as the content of a text file that is
+ being regenerated.
+
presence_penalty: Number between -2.0 and 2.0. Positive values penalize new tokens based on
whether they appear in the text so far, increasing the model's likelihood to
talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
response_format: An object specifying the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4o mini](https://platform.openai.com/docs/models/gpt-4o-mini),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4o mini](https://platform.openai.com/docs/models#gpt-4o-mini),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4) and
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -282,7 +287,7 @@ def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -313,6 +318,7 @@ def create(
modalities: Optional[List[ChatCompletionModality]] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
+ prediction: Optional[ChatCompletionPredictionContentParam] | NotGiven = NOT_GIVEN,
presence_penalty: Optional[float] | NotGiven = NOT_GIVEN,
response_format: completion_create_params.ResponseFormat | NotGiven = NOT_GIVEN,
seed: Optional[int] | NotGiven = NOT_GIVEN,
@@ -349,7 +355,7 @@ def create(
[audio](https://platform.openai.com/docs/guides/audio).
model: ID of the model to use. See the
- [model endpoint compatibility](https://platform.openai.com/docs/models/model-endpoint-compatibility)
+ [model endpoint compatibility](https://platform.openai.com/docs/models#model-endpoint-compatibility)
table for details on which models work with the Chat API.
stream: If set, partial message deltas will be sent, like in ChatGPT. Tokens will be
@@ -367,7 +373,7 @@ def create(
existing frequency in the text so far, decreasing the model's likelihood to
repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
function_call: Deprecated in favor of `tool_choice`.
@@ -428,19 +434,22 @@ def create(
choices. Keep `n` as `1` to minimize costs.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
+ prediction: Static predicted output content, such as the content of a text file that is
+ being regenerated.
+
presence_penalty: Number between -2.0 and 2.0. Positive values penalize new tokens based on
whether they appear in the text so far, increasing the model's likelihood to
talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
response_format: An object specifying the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4o mini](https://platform.openai.com/docs/models/gpt-4o-mini),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4o mini](https://platform.openai.com/docs/models#gpt-4o-mini),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4) and
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -520,7 +529,7 @@ def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -551,6 +560,7 @@ def create(
modalities: Optional[List[ChatCompletionModality]] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
+ prediction: Optional[ChatCompletionPredictionContentParam] | NotGiven = NOT_GIVEN,
presence_penalty: Optional[float] | NotGiven = NOT_GIVEN,
response_format: completion_create_params.ResponseFormat | NotGiven = NOT_GIVEN,
seed: Optional[int] | NotGiven = NOT_GIVEN,
@@ -587,7 +597,7 @@ def create(
[audio](https://platform.openai.com/docs/guides/audio).
model: ID of the model to use. See the
- [model endpoint compatibility](https://platform.openai.com/docs/models/model-endpoint-compatibility)
+ [model endpoint compatibility](https://platform.openai.com/docs/models#model-endpoint-compatibility)
table for details on which models work with the Chat API.
stream: If set, partial message deltas will be sent, like in ChatGPT. Tokens will be
@@ -605,7 +615,7 @@ def create(
existing frequency in the text so far, decreasing the model's likelihood to
repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
function_call: Deprecated in favor of `tool_choice`.
@@ -666,19 +676,22 @@ def create(
choices. Keep `n` as `1` to minimize costs.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
+ prediction: Static predicted output content, such as the content of a text file that is
+ being regenerated.
+
presence_penalty: Number between -2.0 and 2.0. Positive values penalize new tokens based on
whether they appear in the text so far, increasing the model's likelihood to
talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
response_format: An object specifying the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4o mini](https://platform.openai.com/docs/models/gpt-4o-mini),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4o mini](https://platform.openai.com/docs/models#gpt-4o-mini),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4) and
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -758,7 +771,7 @@ def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -788,6 +801,7 @@ def create(
modalities: Optional[List[ChatCompletionModality]] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
+ prediction: Optional[ChatCompletionPredictionContentParam] | NotGiven = NOT_GIVEN,
presence_penalty: Optional[float] | NotGiven = NOT_GIVEN,
response_format: completion_create_params.ResponseFormat | NotGiven = NOT_GIVEN,
seed: Optional[int] | NotGiven = NOT_GIVEN,
@@ -827,6 +841,7 @@ def create(
"modalities": modalities,
"n": n,
"parallel_tool_calls": parallel_tool_calls,
+ "prediction": prediction,
"presence_penalty": presence_penalty,
"response_format": response_format,
"seed": seed,
@@ -891,6 +906,7 @@ async def create(
modalities: Optional[List[ChatCompletionModality]] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
+ prediction: Optional[ChatCompletionPredictionContentParam] | NotGiven = NOT_GIVEN,
presence_penalty: Optional[float] | NotGiven = NOT_GIVEN,
response_format: completion_create_params.ResponseFormat | NotGiven = NOT_GIVEN,
seed: Optional[int] | NotGiven = NOT_GIVEN,
@@ -928,7 +944,7 @@ async def create(
[audio](https://platform.openai.com/docs/guides/audio).
model: ID of the model to use. See the
- [model endpoint compatibility](https://platform.openai.com/docs/models/model-endpoint-compatibility)
+ [model endpoint compatibility](https://platform.openai.com/docs/models#model-endpoint-compatibility)
table for details on which models work with the Chat API.
audio: Parameters for audio output. Required when audio output is requested with
@@ -939,7 +955,7 @@ async def create(
existing frequency in the text so far, decreasing the model's likelihood to
repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
function_call: Deprecated in favor of `tool_choice`.
@@ -1000,19 +1016,22 @@ async def create(
choices. Keep `n` as `1` to minimize costs.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
+ prediction: Static predicted output content, such as the content of a text file that is
+ being regenerated.
+
presence_penalty: Number between -2.0 and 2.0. Positive values penalize new tokens based on
whether they appear in the text so far, increasing the model's likelihood to
talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
response_format: An object specifying the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4o mini](https://platform.openai.com/docs/models/gpt-4o-mini),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4o mini](https://platform.openai.com/docs/models#gpt-4o-mini),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4) and
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -1099,7 +1118,7 @@ async def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -1130,6 +1149,7 @@ async def create(
modalities: Optional[List[ChatCompletionModality]] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
+ prediction: Optional[ChatCompletionPredictionContentParam] | NotGiven = NOT_GIVEN,
presence_penalty: Optional[float] | NotGiven = NOT_GIVEN,
response_format: completion_create_params.ResponseFormat | NotGiven = NOT_GIVEN,
seed: Optional[int] | NotGiven = NOT_GIVEN,
@@ -1166,7 +1186,7 @@ async def create(
[audio](https://platform.openai.com/docs/guides/audio).
model: ID of the model to use. See the
- [model endpoint compatibility](https://platform.openai.com/docs/models/model-endpoint-compatibility)
+ [model endpoint compatibility](https://platform.openai.com/docs/models#model-endpoint-compatibility)
table for details on which models work with the Chat API.
stream: If set, partial message deltas will be sent, like in ChatGPT. Tokens will be
@@ -1184,7 +1204,7 @@ async def create(
existing frequency in the text so far, decreasing the model's likelihood to
repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
function_call: Deprecated in favor of `tool_choice`.
@@ -1245,19 +1265,22 @@ async def create(
choices. Keep `n` as `1` to minimize costs.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
+ prediction: Static predicted output content, such as the content of a text file that is
+ being regenerated.
+
presence_penalty: Number between -2.0 and 2.0. Positive values penalize new tokens based on
whether they appear in the text so far, increasing the model's likelihood to
talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
response_format: An object specifying the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4o mini](https://platform.openai.com/docs/models/gpt-4o-mini),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4o mini](https://platform.openai.com/docs/models#gpt-4o-mini),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4) and
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -1337,7 +1360,7 @@ async def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -1368,6 +1391,7 @@ async def create(
modalities: Optional[List[ChatCompletionModality]] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
+ prediction: Optional[ChatCompletionPredictionContentParam] | NotGiven = NOT_GIVEN,
presence_penalty: Optional[float] | NotGiven = NOT_GIVEN,
response_format: completion_create_params.ResponseFormat | NotGiven = NOT_GIVEN,
seed: Optional[int] | NotGiven = NOT_GIVEN,
@@ -1404,7 +1428,7 @@ async def create(
[audio](https://platform.openai.com/docs/guides/audio).
model: ID of the model to use. See the
- [model endpoint compatibility](https://platform.openai.com/docs/models/model-endpoint-compatibility)
+ [model endpoint compatibility](https://platform.openai.com/docs/models#model-endpoint-compatibility)
table for details on which models work with the Chat API.
stream: If set, partial message deltas will be sent, like in ChatGPT. Tokens will be
@@ -1422,7 +1446,7 @@ async def create(
existing frequency in the text so far, decreasing the model's likelihood to
repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
function_call: Deprecated in favor of `tool_choice`.
@@ -1483,19 +1507,22 @@ async def create(
choices. Keep `n` as `1` to minimize costs.
parallel_tool_calls: Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
+ prediction: Static predicted output content, such as the content of a text file that is
+ being regenerated.
+
presence_penalty: Number between -2.0 and 2.0. Positive values penalize new tokens based on
whether they appear in the text so far, increasing the model's likelihood to
talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
response_format: An object specifying the format that the model must output. Compatible with
- [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4o mini](https://platform.openai.com/docs/models/gpt-4o-mini),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
+ [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4o mini](https://platform.openai.com/docs/models#gpt-4o-mini),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4) and
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -1575,7 +1602,7 @@ async def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -1605,6 +1632,7 @@ async def create(
modalities: Optional[List[ChatCompletionModality]] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
+ prediction: Optional[ChatCompletionPredictionContentParam] | NotGiven = NOT_GIVEN,
presence_penalty: Optional[float] | NotGiven = NOT_GIVEN,
response_format: completion_create_params.ResponseFormat | NotGiven = NOT_GIVEN,
seed: Optional[int] | NotGiven = NOT_GIVEN,
@@ -1644,6 +1672,7 @@ async def create(
"modalities": modalities,
"n": n,
"parallel_tool_calls": parallel_tool_calls,
+ "prediction": prediction,
"presence_penalty": presence_penalty,
"response_format": response_format,
"seed": seed,
diff --git a/src/openai/resources/completions.py b/src/openai/resources/completions.py
index 7f5a3fc4ff..7e95f79607 100644
--- a/src/openai/resources/completions.py
+++ b/src/openai/resources/completions.py
@@ -82,8 +82,8 @@ def create(
model: ID of the model to use. You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
prompt: The prompt(s) to generate completions for, encoded as a string, array of
strings, array of tokens, or array of token arrays.
@@ -108,7 +108,7 @@ def create(
existing frequency in the text so far, decreasing the model's likelihood to
repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
logit_bias: Modify the likelihood of specified tokens appearing in the completion.
@@ -148,7 +148,7 @@ def create(
whether they appear in the text so far, increasing the model's likelihood to
talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
seed: If specified, our system will make a best effort to sample deterministically,
such that repeated requests with the same `seed` and parameters should return
@@ -187,7 +187,7 @@ def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -235,8 +235,8 @@ def create(
model: ID of the model to use. You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
prompt: The prompt(s) to generate completions for, encoded as a string, array of
strings, array of tokens, or array of token arrays.
@@ -268,7 +268,7 @@ def create(
existing frequency in the text so far, decreasing the model's likelihood to
repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
logit_bias: Modify the likelihood of specified tokens appearing in the completion.
@@ -308,7 +308,7 @@ def create(
whether they appear in the text so far, increasing the model's likelihood to
talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
seed: If specified, our system will make a best effort to sample deterministically,
such that repeated requests with the same `seed` and parameters should return
@@ -340,7 +340,7 @@ def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -388,8 +388,8 @@ def create(
model: ID of the model to use. You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
prompt: The prompt(s) to generate completions for, encoded as a string, array of
strings, array of tokens, or array of token arrays.
@@ -421,7 +421,7 @@ def create(
existing frequency in the text so far, decreasing the model's likelihood to
repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
logit_bias: Modify the likelihood of specified tokens appearing in the completion.
@@ -461,7 +461,7 @@ def create(
whether they appear in the text so far, increasing the model's likelihood to
talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
seed: If specified, our system will make a best effort to sample deterministically,
such that repeated requests with the same `seed` and parameters should return
@@ -493,7 +493,7 @@ def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -624,8 +624,8 @@ async def create(
model: ID of the model to use. You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
prompt: The prompt(s) to generate completions for, encoded as a string, array of
strings, array of tokens, or array of token arrays.
@@ -650,7 +650,7 @@ async def create(
existing frequency in the text so far, decreasing the model's likelihood to
repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
logit_bias: Modify the likelihood of specified tokens appearing in the completion.
@@ -690,7 +690,7 @@ async def create(
whether they appear in the text so far, increasing the model's likelihood to
talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
seed: If specified, our system will make a best effort to sample deterministically,
such that repeated requests with the same `seed` and parameters should return
@@ -729,7 +729,7 @@ async def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -777,8 +777,8 @@ async def create(
model: ID of the model to use. You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
prompt: The prompt(s) to generate completions for, encoded as a string, array of
strings, array of tokens, or array of token arrays.
@@ -810,7 +810,7 @@ async def create(
existing frequency in the text so far, decreasing the model's likelihood to
repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
logit_bias: Modify the likelihood of specified tokens appearing in the completion.
@@ -850,7 +850,7 @@ async def create(
whether they appear in the text so far, increasing the model's likelihood to
talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
seed: If specified, our system will make a best effort to sample deterministically,
such that repeated requests with the same `seed` and parameters should return
@@ -882,7 +882,7 @@ async def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -930,8 +930,8 @@ async def create(
model: ID of the model to use. You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
prompt: The prompt(s) to generate completions for, encoded as a string, array of
strings, array of tokens, or array of token arrays.
@@ -963,7 +963,7 @@ async def create(
existing frequency in the text so far, decreasing the model's likelihood to
repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
logit_bias: Modify the likelihood of specified tokens appearing in the completion.
@@ -1003,7 +1003,7 @@ async def create(
whether they appear in the text so far, increasing the model's likelihood to
talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
seed: If specified, our system will make a best effort to sample deterministically,
such that repeated requests with the same `seed` and parameters should return
@@ -1035,7 +1035,7 @@ async def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
diff --git a/src/openai/resources/embeddings.py b/src/openai/resources/embeddings.py
index 3a2763904b..2197c4d280 100644
--- a/src/openai/resources/embeddings.py
+++ b/src/openai/resources/embeddings.py
@@ -74,8 +74,8 @@ def create(
model: ID of the model to use. You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
dimensions: The number of dimensions the resulting output embeddings should have. Only
supported in `text-embedding-3` and later models.
@@ -85,7 +85,7 @@ def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -164,8 +164,8 @@ async def create(
model: ID of the model to use. You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
dimensions: The number of dimensions the resulting output embeddings should have. Only
supported in `text-embedding-3` and later models.
@@ -175,7 +175,7 @@ async def create(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
diff --git a/src/openai/resources/files.py b/src/openai/resources/files.py
index ee668e9bc2..977b9b3c48 100644
--- a/src/openai/resources/files.py
+++ b/src/openai/resources/files.py
@@ -4,6 +4,7 @@
import typing_extensions
from typing import Mapping, cast
+from typing_extensions import Literal
import httpx
@@ -26,7 +27,7 @@
to_custom_streamed_response_wrapper,
async_to_custom_streamed_response_wrapper,
)
-from ..pagination import SyncPage, AsyncPage
+from ..pagination import SyncCursorPage, AsyncCursorPage
from .._base_client import AsyncPaginator, make_request_options
from ..types.file_object import FileObject
from ..types.file_deleted import FileDeleted
@@ -168,6 +169,9 @@ def retrieve(
def list(
self,
*,
+ after: str | NotGiven = NOT_GIVEN,
+ limit: int | NotGiven = NOT_GIVEN,
+ order: Literal["asc", "desc"] | NotGiven = NOT_GIVEN,
purpose: str | NotGiven = NOT_GIVEN,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
@@ -175,11 +179,23 @@ def list(
extra_query: Query | None = None,
extra_body: Body | None = None,
timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN,
- ) -> SyncPage[FileObject]:
- """
- Returns a list of files that belong to the user's organization.
+ ) -> SyncCursorPage[FileObject]:
+ """Returns a list of files.
Args:
+ after: A cursor for use in pagination.
+
+ `after` is an object ID that defines your place
+ in the list. For instance, if you make a list request and receive 100 objects,
+ ending with obj_foo, your subsequent call can include after=obj_foo in order to
+ fetch the next page of the list.
+
+ limit: A limit on the number of objects to be returned. Limit can range between 1 and
+ 10,000, and the default is 10,000.
+
+ order: Sort order by the `created_at` timestamp of the objects. `asc` for ascending
+ order and `desc` for descending order.
+
purpose: Only return files with the given purpose.
extra_headers: Send extra headers
@@ -192,13 +208,21 @@ def list(
"""
return self._get_api_list(
"/files",
- page=SyncPage[FileObject],
+ page=SyncCursorPage[FileObject],
options=make_request_options(
extra_headers=extra_headers,
extra_query=extra_query,
extra_body=extra_body,
timeout=timeout,
- query=maybe_transform({"purpose": purpose}, file_list_params.FileListParams),
+ query=maybe_transform(
+ {
+ "after": after,
+ "limit": limit,
+ "order": order,
+ "purpose": purpose,
+ },
+ file_list_params.FileListParams,
+ ),
),
model=FileObject,
)
@@ -438,6 +462,9 @@ async def retrieve(
def list(
self,
*,
+ after: str | NotGiven = NOT_GIVEN,
+ limit: int | NotGiven = NOT_GIVEN,
+ order: Literal["asc", "desc"] | NotGiven = NOT_GIVEN,
purpose: str | NotGiven = NOT_GIVEN,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
@@ -445,11 +472,23 @@ def list(
extra_query: Query | None = None,
extra_body: Body | None = None,
timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN,
- ) -> AsyncPaginator[FileObject, AsyncPage[FileObject]]:
- """
- Returns a list of files that belong to the user's organization.
+ ) -> AsyncPaginator[FileObject, AsyncCursorPage[FileObject]]:
+ """Returns a list of files.
Args:
+ after: A cursor for use in pagination.
+
+ `after` is an object ID that defines your place
+ in the list. For instance, if you make a list request and receive 100 objects,
+ ending with obj_foo, your subsequent call can include after=obj_foo in order to
+ fetch the next page of the list.
+
+ limit: A limit on the number of objects to be returned. Limit can range between 1 and
+ 10,000, and the default is 10,000.
+
+ order: Sort order by the `created_at` timestamp of the objects. `asc` for ascending
+ order and `desc` for descending order.
+
purpose: Only return files with the given purpose.
extra_headers: Send extra headers
@@ -462,13 +501,21 @@ def list(
"""
return self._get_api_list(
"/files",
- page=AsyncPage[FileObject],
+ page=AsyncCursorPage[FileObject],
options=make_request_options(
extra_headers=extra_headers,
extra_query=extra_query,
extra_body=extra_body,
timeout=timeout,
- query=maybe_transform({"purpose": purpose}, file_list_params.FileListParams),
+ query=maybe_transform(
+ {
+ "after": after,
+ "limit": limit,
+ "order": order,
+ "purpose": purpose,
+ },
+ file_list_params.FileListParams,
+ ),
),
model=FileObject,
)
diff --git a/src/openai/resources/fine_tuning/jobs/jobs.py b/src/openai/resources/fine_tuning/jobs/jobs.py
index 7eb0c5dbfc..4024bf79f3 100644
--- a/src/openai/resources/fine_tuning/jobs/jobs.py
+++ b/src/openai/resources/fine_tuning/jobs/jobs.py
@@ -85,7 +85,7 @@ def create(
Args:
model: The name of the model to fine-tune. You can select one of the
- [supported models](https://platform.openai.com/docs/guides/fine-tuning/which-models-can-be-fine-tuned).
+ [supported models](https://platform.openai.com/docs/guides/fine-tuning#which-models-can-be-fine-tuned).
training_file: The ID of an uploaded file that contains training data.
@@ -376,7 +376,7 @@ async def create(
Args:
model: The name of the model to fine-tune. You can select one of the
- [supported models](https://platform.openai.com/docs/guides/fine-tuning/which-models-can-be-fine-tuned).
+ [supported models](https://platform.openai.com/docs/guides/fine-tuning#which-models-can-be-fine-tuned).
training_file: The ID of an uploaded file that contains training data.
diff --git a/src/openai/resources/images.py b/src/openai/resources/images.py
index e9629d48fd..2fbc077dd9 100644
--- a/src/openai/resources/images.py
+++ b/src/openai/resources/images.py
@@ -84,7 +84,7 @@ def create_variation(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -165,7 +165,7 @@ def edit(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -251,7 +251,7 @@ def generate(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -341,7 +341,7 @@ async def create_variation(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -422,7 +422,7 @@ async def edit(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
@@ -508,7 +508,7 @@ async def generate(
user: A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
extra_headers: Send extra headers
diff --git a/src/openai/resources/moderations.py b/src/openai/resources/moderations.py
index 8b73da57b2..ce80bb7d55 100644
--- a/src/openai/resources/moderations.py
+++ b/src/openai/resources/moderations.py
@@ -68,7 +68,7 @@ def create(
model: The content moderation model you would like to use. Learn more in
[the moderation guide](https://platform.openai.com/docs/guides/moderation), and
learn about available models
- [here](https://platform.openai.com/docs/models/moderation).
+ [here](https://platform.openai.com/docs/models#moderation).
extra_headers: Send extra headers
@@ -138,7 +138,7 @@ async def create(
model: The content moderation model you would like to use. Learn more in
[the moderation guide](https://platform.openai.com/docs/guides/moderation), and
learn about available models
- [here](https://platform.openai.com/docs/models/moderation).
+ [here](https://platform.openai.com/docs/models#moderation).
extra_headers: Send extra headers
diff --git a/src/openai/resources/uploads/uploads.py b/src/openai/resources/uploads/uploads.py
index dadd01d342..2384716bdd 100644
--- a/src/openai/resources/uploads/uploads.py
+++ b/src/openai/resources/uploads/uploads.py
@@ -85,7 +85,7 @@ def create(
For certain `purpose`s, the correct `mime_type` must be specified. Please refer
to documentation for the supported MIME types for your use case:
- - [Assistants](https://platform.openai.com/docs/assistants/tools/file-search/supported-files)
+ - [Assistants](https://platform.openai.com/docs/assistants/tools/file-search#supported-files)
For guidance on the proper filename extensions for each purpose, please follow
the documentation on
@@ -279,7 +279,7 @@ async def create(
For certain `purpose`s, the correct `mime_type` must be specified. Please refer
to documentation for the supported MIME types for your use case:
- - [Assistants](https://platform.openai.com/docs/assistants/tools/file-search/supported-files)
+ - [Assistants](https://platform.openai.com/docs/assistants/tools/file-search#supported-files)
For guidance on the proper filename extensions for each purpose, please follow
the documentation on
diff --git a/src/openai/types/audio/speech_create_params.py b/src/openai/types/audio/speech_create_params.py
index dff66e49c7..a60d000708 100644
--- a/src/openai/types/audio/speech_create_params.py
+++ b/src/openai/types/audio/speech_create_params.py
@@ -16,7 +16,7 @@ class SpeechCreateParams(TypedDict, total=False):
model: Required[Union[str, SpeechModel]]
"""
- One of the available [TTS models](https://platform.openai.com/docs/models/tts):
+ One of the available [TTS models](https://platform.openai.com/docs/models#tts):
`tts-1` or `tts-1-hd`
"""
@@ -25,7 +25,7 @@ class SpeechCreateParams(TypedDict, total=False):
Supported voices are `alloy`, `echo`, `fable`, `onyx`, `nova`, and `shimmer`.
Previews of the voices are available in the
- [Text to speech guide](https://platform.openai.com/docs/guides/text-to-speech/voice-options).
+ [Text to speech guide](https://platform.openai.com/docs/guides/text-to-speech#voice-options).
"""
response_format: Literal["mp3", "opus", "aac", "flac", "wav", "pcm"]
diff --git a/src/openai/types/audio/transcription_create_params.py b/src/openai/types/audio/transcription_create_params.py
index 5ac2bb91e5..88805affbd 100644
--- a/src/openai/types/audio/transcription_create_params.py
+++ b/src/openai/types/audio/transcription_create_params.py
@@ -38,7 +38,7 @@ class TranscriptionCreateParams(TypedDict, total=False):
"""An optional text to guide the model's style or continue a previous audio
segment.
- The [prompt](https://platform.openai.com/docs/guides/speech-to-text/prompting)
+ The [prompt](https://platform.openai.com/docs/guides/speech-to-text#prompting)
should match the audio language.
"""
diff --git a/src/openai/types/audio/translation_create_params.py b/src/openai/types/audio/translation_create_params.py
index 6859ed9d30..62f85b8757 100644
--- a/src/openai/types/audio/translation_create_params.py
+++ b/src/openai/types/audio/translation_create_params.py
@@ -30,7 +30,7 @@ class TranslationCreateParams(TypedDict, total=False):
"""An optional text to guide the model's style or continue a previous audio
segment.
- The [prompt](https://platform.openai.com/docs/guides/speech-to-text/prompting)
+ The [prompt](https://platform.openai.com/docs/guides/speech-to-text#prompting)
should be in English.
"""
diff --git a/src/openai/types/beta/assistant.py b/src/openai/types/beta/assistant.py
index ea97de440f..3c8b8e403b 100644
--- a/src/openai/types/beta/assistant.py
+++ b/src/openai/types/beta/assistant.py
@@ -65,8 +65,8 @@ class Assistant(BaseModel):
You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
"""
name: Optional[str] = None
@@ -85,8 +85,8 @@ class Assistant(BaseModel):
response_format: Optional[AssistantResponseFormatOption] = None
"""Specifies the format that the model must output.
- Compatible with [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
diff --git a/src/openai/types/beta/assistant_create_params.py b/src/openai/types/beta/assistant_create_params.py
index e11f842f05..568b223ce7 100644
--- a/src/openai/types/beta/assistant_create_params.py
+++ b/src/openai/types/beta/assistant_create_params.py
@@ -26,8 +26,8 @@ class AssistantCreateParams(TypedDict, total=False):
You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
"""
description: Optional[str]
@@ -53,8 +53,8 @@ class AssistantCreateParams(TypedDict, total=False):
response_format: Optional[AssistantResponseFormatOptionParam]
"""Specifies the format that the model must output.
- Compatible with [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
diff --git a/src/openai/types/beta/assistant_list_params.py b/src/openai/types/beta/assistant_list_params.py
index f54f63120b..834ffbcaf8 100644
--- a/src/openai/types/beta/assistant_list_params.py
+++ b/src/openai/types/beta/assistant_list_params.py
@@ -21,7 +21,7 @@ class AssistantListParams(TypedDict, total=False):
"""A cursor for use in pagination.
`before` is an object ID that defines your place in the list. For instance, if
- you make a list request and receive 100 objects, ending with obj_foo, your
+ you make a list request and receive 100 objects, starting with obj_foo, your
subsequent call can include before=obj_foo in order to fetch the previous page
of the list.
"""
diff --git a/src/openai/types/beta/assistant_update_params.py b/src/openai/types/beta/assistant_update_params.py
index c4598df507..9a66e41ab3 100644
--- a/src/openai/types/beta/assistant_update_params.py
+++ b/src/openai/types/beta/assistant_update_params.py
@@ -35,8 +35,8 @@ class AssistantUpdateParams(TypedDict, total=False):
You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
"""
name: Optional[str]
@@ -45,8 +45,8 @@ class AssistantUpdateParams(TypedDict, total=False):
response_format: Optional[AssistantResponseFormatOptionParam]
"""Specifies the format that the model must output.
- Compatible with [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
diff --git a/src/openai/types/beta/file_search_tool.py b/src/openai/types/beta/file_search_tool.py
index aee6593e89..89fc16c04c 100644
--- a/src/openai/types/beta/file_search_tool.py
+++ b/src/openai/types/beta/file_search_tool.py
@@ -31,7 +31,7 @@ class FileSearch(BaseModel):
Note that the file search tool may output fewer than `max_num_results` results.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
"""
@@ -42,7 +42,7 @@ class FileSearch(BaseModel):
score_threshold of 0.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
"""
diff --git a/src/openai/types/beta/file_search_tool_param.py b/src/openai/types/beta/file_search_tool_param.py
index 5ce91207ba..c73d0af79d 100644
--- a/src/openai/types/beta/file_search_tool_param.py
+++ b/src/openai/types/beta/file_search_tool_param.py
@@ -30,7 +30,7 @@ class FileSearch(TypedDict, total=False):
Note that the file search tool may output fewer than `max_num_results` results.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
"""
@@ -41,7 +41,7 @@ class FileSearch(TypedDict, total=False):
score_threshold of 0.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
"""
diff --git a/src/openai/types/beta/thread_create_and_run_params.py b/src/openai/types/beta/thread_create_and_run_params.py
index 64ee6a8710..8310ba12f4 100644
--- a/src/openai/types/beta/thread_create_and_run_params.py
+++ b/src/openai/types/beta/thread_create_and_run_params.py
@@ -86,15 +86,15 @@ class ThreadCreateAndRunParamsBase(TypedDict, total=False):
parallel_tool_calls: bool
"""
Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
"""
response_format: Optional[AssistantResponseFormatOptionParam]
"""Specifies the format that the model must output.
- Compatible with [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
diff --git a/src/openai/types/beta/threads/message_list_params.py b/src/openai/types/beta/threads/message_list_params.py
index 18c2442fb5..a7c22a66fb 100644
--- a/src/openai/types/beta/threads/message_list_params.py
+++ b/src/openai/types/beta/threads/message_list_params.py
@@ -21,7 +21,7 @@ class MessageListParams(TypedDict, total=False):
"""A cursor for use in pagination.
`before` is an object ID that defines your place in the list. For instance, if
- you make a list request and receive 100 objects, ending with obj_foo, your
+ you make a list request and receive 100 objects, starting with obj_foo, your
subsequent call can include before=obj_foo in order to fetch the previous page
of the list.
"""
diff --git a/src/openai/types/beta/threads/run.py b/src/openai/types/beta/threads/run.py
index e8f2b74dee..ad32135b7d 100644
--- a/src/openai/types/beta/threads/run.py
+++ b/src/openai/types/beta/threads/run.py
@@ -154,7 +154,7 @@ class Run(BaseModel):
parallel_tool_calls: bool
"""
Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
"""
@@ -167,8 +167,8 @@ class Run(BaseModel):
response_format: Optional[AssistantResponseFormatOption] = None
"""Specifies the format that the model must output.
- Compatible with [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
diff --git a/src/openai/types/beta/threads/run_create_params.py b/src/openai/types/beta/threads/run_create_params.py
index 9767b142e1..88dc39645e 100644
--- a/src/openai/types/beta/threads/run_create_params.py
+++ b/src/openai/types/beta/threads/run_create_params.py
@@ -41,7 +41,7 @@ class RunCreateParamsBase(TypedDict, total=False):
search result content.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
"""
@@ -99,15 +99,15 @@ class RunCreateParamsBase(TypedDict, total=False):
parallel_tool_calls: bool
"""
Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
"""
response_format: Optional[AssistantResponseFormatOptionParam]
"""Specifies the format that the model must output.
- Compatible with [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
diff --git a/src/openai/types/beta/threads/run_list_params.py b/src/openai/types/beta/threads/run_list_params.py
index 1e32bca4b4..fbea54f6f2 100644
--- a/src/openai/types/beta/threads/run_list_params.py
+++ b/src/openai/types/beta/threads/run_list_params.py
@@ -21,7 +21,7 @@ class RunListParams(TypedDict, total=False):
"""A cursor for use in pagination.
`before` is an object ID that defines your place in the list. For instance, if
- you make a list request and receive 100 objects, ending with obj_foo, your
+ you make a list request and receive 100 objects, starting with obj_foo, your
subsequent call can include before=obj_foo in order to fetch the previous page
of the list.
"""
diff --git a/src/openai/types/beta/threads/runs/step_list_params.py b/src/openai/types/beta/threads/runs/step_list_params.py
index 3931bd7e0c..a6be771d9f 100644
--- a/src/openai/types/beta/threads/runs/step_list_params.py
+++ b/src/openai/types/beta/threads/runs/step_list_params.py
@@ -26,7 +26,7 @@ class StepListParams(TypedDict, total=False):
"""A cursor for use in pagination.
`before` is an object ID that defines your place in the list. For instance, if
- you make a list request and receive 100 objects, ending with obj_foo, your
+ you make a list request and receive 100 objects, starting with obj_foo, your
subsequent call can include before=obj_foo in order to fetch the previous page
of the list.
"""
@@ -39,7 +39,7 @@ class StepListParams(TypedDict, total=False):
search result content.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
"""
diff --git a/src/openai/types/beta/threads/runs/step_retrieve_params.py b/src/openai/types/beta/threads/runs/step_retrieve_params.py
index 22c1c049f4..ecbb72edbd 100644
--- a/src/openai/types/beta/threads/runs/step_retrieve_params.py
+++ b/src/openai/types/beta/threads/runs/step_retrieve_params.py
@@ -23,6 +23,6 @@ class StepRetrieveParams(TypedDict, total=False):
search result content.
See the
- [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
+ [file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search#customizing-file-search-settings)
for more information.
"""
diff --git a/src/openai/types/beta/vector_store_list_params.py b/src/openai/types/beta/vector_store_list_params.py
index f39f67266d..e26ff90a85 100644
--- a/src/openai/types/beta/vector_store_list_params.py
+++ b/src/openai/types/beta/vector_store_list_params.py
@@ -21,7 +21,7 @@ class VectorStoreListParams(TypedDict, total=False):
"""A cursor for use in pagination.
`before` is an object ID that defines your place in the list. For instance, if
- you make a list request and receive 100 objects, ending with obj_foo, your
+ you make a list request and receive 100 objects, starting with obj_foo, your
subsequent call can include before=obj_foo in order to fetch the previous page
of the list.
"""
diff --git a/src/openai/types/beta/vector_stores/file_batch_list_files_params.py b/src/openai/types/beta/vector_stores/file_batch_list_files_params.py
index 24dee7d5a5..2a0a6c6aa7 100644
--- a/src/openai/types/beta/vector_stores/file_batch_list_files_params.py
+++ b/src/openai/types/beta/vector_stores/file_batch_list_files_params.py
@@ -23,7 +23,7 @@ class FileBatchListFilesParams(TypedDict, total=False):
"""A cursor for use in pagination.
`before` is an object ID that defines your place in the list. For instance, if
- you make a list request and receive 100 objects, ending with obj_foo, your
+ you make a list request and receive 100 objects, starting with obj_foo, your
subsequent call can include before=obj_foo in order to fetch the previous page
of the list.
"""
diff --git a/src/openai/types/beta/vector_stores/file_list_params.py b/src/openai/types/beta/vector_stores/file_list_params.py
index 23dd7f0d94..867b5fb3bb 100644
--- a/src/openai/types/beta/vector_stores/file_list_params.py
+++ b/src/openai/types/beta/vector_stores/file_list_params.py
@@ -21,7 +21,7 @@ class FileListParams(TypedDict, total=False):
"""A cursor for use in pagination.
`before` is an object ID that defines your place in the list. For instance, if
- you make a list request and receive 100 objects, ending with obj_foo, your
+ you make a list request and receive 100 objects, starting with obj_foo, your
subsequent call can include before=obj_foo in order to fetch the previous page
of the list.
"""
diff --git a/src/openai/types/chat/__init__.py b/src/openai/types/chat/__init__.py
index b85365ecb1..ef562a4b94 100644
--- a/src/openai/types/chat/__init__.py
+++ b/src/openai/types/chat/__init__.py
@@ -37,6 +37,9 @@
from .chat_completion_content_part_image_param import (
ChatCompletionContentPartImageParam as ChatCompletionContentPartImageParam,
)
+from .chat_completion_prediction_content_param import (
+ ChatCompletionPredictionContentParam as ChatCompletionPredictionContentParam,
+)
from .chat_completion_tool_choice_option_param import (
ChatCompletionToolChoiceOptionParam as ChatCompletionToolChoiceOptionParam,
)
diff --git a/src/openai/types/chat/chat_completion_content_part_image_param.py b/src/openai/types/chat/chat_completion_content_part_image_param.py
index b1a186aa6d..9d407324d0 100644
--- a/src/openai/types/chat/chat_completion_content_part_image_param.py
+++ b/src/openai/types/chat/chat_completion_content_part_image_param.py
@@ -15,7 +15,7 @@ class ImageURL(TypedDict, total=False):
"""Specifies the detail level of the image.
Learn more in the
- [Vision guide](https://platform.openai.com/docs/guides/vision/low-or-high-fidelity-image-understanding).
+ [Vision guide](https://platform.openai.com/docs/guides/vision#low-or-high-fidelity-image-understanding).
"""
diff --git a/src/openai/types/chat/chat_completion_prediction_content_param.py b/src/openai/types/chat/chat_completion_prediction_content_param.py
new file mode 100644
index 0000000000..c44e6e3653
--- /dev/null
+++ b/src/openai/types/chat/chat_completion_prediction_content_param.py
@@ -0,0 +1,25 @@
+# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
+
+from __future__ import annotations
+
+from typing import Union, Iterable
+from typing_extensions import Literal, Required, TypedDict
+
+from .chat_completion_content_part_text_param import ChatCompletionContentPartTextParam
+
+__all__ = ["ChatCompletionPredictionContentParam"]
+
+
+class ChatCompletionPredictionContentParam(TypedDict, total=False):
+ content: Required[Union[str, Iterable[ChatCompletionContentPartTextParam]]]
+ """
+ The content that should be matched when generating a model response. If
+ generated tokens would match this content, the entire model response can be
+ returned much more quickly.
+ """
+
+ type: Required[Literal["content"]]
+ """The type of the predicted content you want to provide.
+
+ This type is currently always `content`.
+ """
diff --git a/src/openai/types/chat/completion_create_params.py b/src/openai/types/chat/completion_create_params.py
index af6a47c219..e838858314 100644
--- a/src/openai/types/chat/completion_create_params.py
+++ b/src/openai/types/chat/completion_create_params.py
@@ -13,6 +13,7 @@
from ..shared_params.function_parameters import FunctionParameters
from ..shared_params.response_format_text import ResponseFormatText
from .chat_completion_stream_options_param import ChatCompletionStreamOptionsParam
+from .chat_completion_prediction_content_param import ChatCompletionPredictionContentParam
from .chat_completion_tool_choice_option_param import ChatCompletionToolChoiceOptionParam
from ..shared_params.response_format_json_object import ResponseFormatJSONObject
from ..shared_params.response_format_json_schema import ResponseFormatJSONSchema
@@ -43,7 +44,7 @@ class CompletionCreateParamsBase(TypedDict, total=False):
"""ID of the model to use.
See the
- [model endpoint compatibility](https://platform.openai.com/docs/models/model-endpoint-compatibility)
+ [model endpoint compatibility](https://platform.openai.com/docs/models#model-endpoint-compatibility)
table for details on which models work with the Chat API.
"""
@@ -60,7 +61,7 @@ class CompletionCreateParamsBase(TypedDict, total=False):
Positive values penalize new tokens based on their existing frequency in the
text so far, decreasing the model's likelihood to repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
"""
function_call: FunctionCall
@@ -148,25 +149,31 @@ class CompletionCreateParamsBase(TypedDict, total=False):
parallel_tool_calls: bool
"""
Whether to enable
- [parallel function calling](https://platform.openai.com/docs/guides/function-calling/parallel-function-calling)
+ [parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
during tool use.
"""
+ prediction: Optional[ChatCompletionPredictionContentParam]
+ """
+ Static predicted output content, such as the content of a text file that is
+ being regenerated.
+ """
+
presence_penalty: Optional[float]
"""Number between -2.0 and 2.0.
Positive values penalize new tokens based on whether they appear in the text so
far, increasing the model's likelihood to talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
"""
response_format: ResponseFormat
"""An object specifying the format that the model must output.
- Compatible with [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
- [GPT-4o mini](https://platform.openai.com/docs/models/gpt-4o-mini),
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
+ Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
+ [GPT-4o mini](https://platform.openai.com/docs/models#gpt-4o-mini),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4) and
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -276,7 +283,7 @@ class CompletionCreateParamsBase(TypedDict, total=False):
"""
A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
"""
diff --git a/src/openai/types/completion_create_params.py b/src/openai/types/completion_create_params.py
index 6c112b3902..fdb1680d26 100644
--- a/src/openai/types/completion_create_params.py
+++ b/src/openai/types/completion_create_params.py
@@ -17,8 +17,8 @@ class CompletionCreateParamsBase(TypedDict, total=False):
You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
"""
prompt: Required[Union[str, List[str], Iterable[int], Iterable[Iterable[int]], None]]
@@ -53,7 +53,7 @@ class CompletionCreateParamsBase(TypedDict, total=False):
Positive values penalize new tokens based on their existing frequency in the
text so far, decreasing the model's likelihood to repeat the same line verbatim.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
"""
logit_bias: Optional[Dict[str, int]]
@@ -106,7 +106,7 @@ class CompletionCreateParamsBase(TypedDict, total=False):
Positive values penalize new tokens based on whether they appear in the text so
far, increasing the model's likelihood to talk about new topics.
- [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation/parameter-details)
+ [See more information about frequency and presence penalties.](https://platform.openai.com/docs/guides/text-generation)
"""
seed: Optional[int]
@@ -156,7 +156,7 @@ class CompletionCreateParamsBase(TypedDict, total=False):
"""
A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
"""
diff --git a/src/openai/types/completion_usage.py b/src/openai/types/completion_usage.py
index fe112833e0..d8c4e84cf7 100644
--- a/src/openai/types/completion_usage.py
+++ b/src/openai/types/completion_usage.py
@@ -8,12 +8,26 @@
class CompletionTokensDetails(BaseModel):
+ accepted_prediction_tokens: Optional[int] = None
+ """
+ When using Predicted Outputs, the number of tokens in the prediction that
+ appeared in the completion.
+ """
+
audio_tokens: Optional[int] = None
"""Audio input tokens generated by the model."""
reasoning_tokens: Optional[int] = None
"""Tokens generated by the model for reasoning."""
+ rejected_prediction_tokens: Optional[int] = None
+ """
+ When using Predicted Outputs, the number of tokens in the prediction that did
+ not appear in the completion. However, like reasoning tokens, these tokens are
+ still counted in the total completion tokens for purposes of billing, output,
+ and context window limits.
+ """
+
class PromptTokensDetails(BaseModel):
audio_tokens: Optional[int] = None
diff --git a/src/openai/types/embedding_create_params.py b/src/openai/types/embedding_create_params.py
index 1548cdbd77..1385762885 100644
--- a/src/openai/types/embedding_create_params.py
+++ b/src/openai/types/embedding_create_params.py
@@ -28,8 +28,8 @@ class EmbeddingCreateParams(TypedDict, total=False):
You can use the
[List models](https://platform.openai.com/docs/api-reference/models/list) API to
see all of your available models, or see our
- [Model overview](https://platform.openai.com/docs/models/overview) for
- descriptions of them.
+ [Model overview](https://platform.openai.com/docs/models) for descriptions of
+ them.
"""
dimensions: int
@@ -48,5 +48,5 @@ class EmbeddingCreateParams(TypedDict, total=False):
"""
A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
"""
diff --git a/src/openai/types/file_list_params.py b/src/openai/types/file_list_params.py
index 212eca13c0..058d874c29 100644
--- a/src/openai/types/file_list_params.py
+++ b/src/openai/types/file_list_params.py
@@ -2,11 +2,32 @@
from __future__ import annotations
-from typing_extensions import TypedDict
+from typing_extensions import Literal, TypedDict
__all__ = ["FileListParams"]
class FileListParams(TypedDict, total=False):
+ after: str
+ """A cursor for use in pagination.
+
+ `after` is an object ID that defines your place in the list. For instance, if
+ you make a list request and receive 100 objects, ending with obj_foo, your
+ subsequent call can include after=obj_foo in order to fetch the next page of the
+ list.
+ """
+
+ limit: int
+ """A limit on the number of objects to be returned.
+
+ Limit can range between 1 and 10,000, and the default is 10,000.
+ """
+
+ order: Literal["asc", "desc"]
+ """Sort order by the `created_at` timestamp of the objects.
+
+ `asc` for ascending order and `desc` for descending order.
+ """
+
purpose: str
"""Only return files with the given purpose."""
diff --git a/src/openai/types/fine_tuning/job_create_params.py b/src/openai/types/fine_tuning/job_create_params.py
index 8f5ea86274..8814229b2e 100644
--- a/src/openai/types/fine_tuning/job_create_params.py
+++ b/src/openai/types/fine_tuning/job_create_params.py
@@ -13,7 +13,7 @@ class JobCreateParams(TypedDict, total=False):
"""The name of the model to fine-tune.
You can select one of the
- [supported models](https://platform.openai.com/docs/guides/fine-tuning/which-models-can-be-fine-tuned).
+ [supported models](https://platform.openai.com/docs/guides/fine-tuning#which-models-can-be-fine-tuned).
"""
training_file: Required[str]
diff --git a/src/openai/types/image_create_variation_params.py b/src/openai/types/image_create_variation_params.py
index d6ecf0f1ae..d20f672912 100644
--- a/src/openai/types/image_create_variation_params.py
+++ b/src/openai/types/image_create_variation_params.py
@@ -47,5 +47,5 @@ class ImageCreateVariationParams(TypedDict, total=False):
"""
A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
"""
diff --git a/src/openai/types/image_edit_params.py b/src/openai/types/image_edit_params.py
index a596a8692b..1cb10611f3 100644
--- a/src/openai/types/image_edit_params.py
+++ b/src/openai/types/image_edit_params.py
@@ -58,5 +58,5 @@ class ImageEditParams(TypedDict, total=False):
"""
A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
"""
diff --git a/src/openai/types/image_generate_params.py b/src/openai/types/image_generate_params.py
index 307adeb3da..c88c45f518 100644
--- a/src/openai/types/image_generate_params.py
+++ b/src/openai/types/image_generate_params.py
@@ -61,5 +61,5 @@ class ImageGenerateParams(TypedDict, total=False):
"""
A unique identifier representing your end-user, which can help OpenAI to monitor
and detect abuse.
- [Learn more](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids).
+ [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids).
"""
diff --git a/src/openai/types/moderation_create_params.py b/src/openai/types/moderation_create_params.py
index 3193fd9c2d..3ea2f3cd88 100644
--- a/src/openai/types/moderation_create_params.py
+++ b/src/openai/types/moderation_create_params.py
@@ -25,5 +25,5 @@ class ModerationCreateParams(TypedDict, total=False):
Learn more in
[the moderation guide](https://platform.openai.com/docs/guides/moderation), and
learn about available models
- [here](https://platform.openai.com/docs/models/moderation).
+ [here](https://platform.openai.com/docs/models#moderation).
"""
diff --git a/tests/api_resources/chat/test_completions.py b/tests/api_resources/chat/test_completions.py
index a341e78f7e..4359993e9c 100644
--- a/tests/api_resources/chat/test_completions.py
+++ b/tests/api_resources/chat/test_completions.py
@@ -64,6 +64,10 @@ def test_method_create_with_all_params_overload_1(self, client: OpenAI) -> None:
modalities=["text", "audio"],
n=1,
parallel_tool_calls=True,
+ prediction={
+ "content": "string",
+ "type": "content",
+ },
presence_penalty=-2,
response_format={"type": "text"},
seed=-9007199254740991,
@@ -192,6 +196,10 @@ def test_method_create_with_all_params_overload_2(self, client: OpenAI) -> None:
modalities=["text", "audio"],
n=1,
parallel_tool_calls=True,
+ prediction={
+ "content": "string",
+ "type": "content",
+ },
presence_penalty=-2,
response_format={"type": "text"},
seed=-9007199254740991,
@@ -322,6 +330,10 @@ async def test_method_create_with_all_params_overload_1(self, async_client: Asyn
modalities=["text", "audio"],
n=1,
parallel_tool_calls=True,
+ prediction={
+ "content": "string",
+ "type": "content",
+ },
presence_penalty=-2,
response_format={"type": "text"},
seed=-9007199254740991,
@@ -450,6 +462,10 @@ async def test_method_create_with_all_params_overload_2(self, async_client: Asyn
modalities=["text", "audio"],
n=1,
parallel_tool_calls=True,
+ prediction={
+ "content": "string",
+ "type": "content",
+ },
presence_penalty=-2,
response_format={"type": "text"},
seed=-9007199254740991,
diff --git a/tests/api_resources/test_files.py b/tests/api_resources/test_files.py
index 725e55c193..62d2b88ebf 100644
--- a/tests/api_resources/test_files.py
+++ b/tests/api_resources/test_files.py
@@ -13,7 +13,7 @@
from openai import OpenAI, AsyncOpenAI
from tests.utils import assert_matches_type
from openai.types import FileObject, FileDeleted
-from openai.pagination import SyncPage, AsyncPage
+from openai.pagination import SyncCursorPage, AsyncCursorPage
# pyright: reportDeprecated=false
@@ -98,14 +98,17 @@ def test_path_params_retrieve(self, client: OpenAI) -> None:
@parametrize
def test_method_list(self, client: OpenAI) -> None:
file = client.files.list()
- assert_matches_type(SyncPage[FileObject], file, path=["response"])
+ assert_matches_type(SyncCursorPage[FileObject], file, path=["response"])
@parametrize
def test_method_list_with_all_params(self, client: OpenAI) -> None:
file = client.files.list(
+ after="after",
+ limit=0,
+ order="asc",
purpose="purpose",
)
- assert_matches_type(SyncPage[FileObject], file, path=["response"])
+ assert_matches_type(SyncCursorPage[FileObject], file, path=["response"])
@parametrize
def test_raw_response_list(self, client: OpenAI) -> None:
@@ -114,7 +117,7 @@ def test_raw_response_list(self, client: OpenAI) -> None:
assert response.is_closed is True
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
file = response.parse()
- assert_matches_type(SyncPage[FileObject], file, path=["response"])
+ assert_matches_type(SyncCursorPage[FileObject], file, path=["response"])
@parametrize
def test_streaming_response_list(self, client: OpenAI) -> None:
@@ -123,7 +126,7 @@ def test_streaming_response_list(self, client: OpenAI) -> None:
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
file = response.parse()
- assert_matches_type(SyncPage[FileObject], file, path=["response"])
+ assert_matches_type(SyncCursorPage[FileObject], file, path=["response"])
assert cast(Any, response.is_closed) is True
@@ -334,14 +337,17 @@ async def test_path_params_retrieve(self, async_client: AsyncOpenAI) -> None:
@parametrize
async def test_method_list(self, async_client: AsyncOpenAI) -> None:
file = await async_client.files.list()
- assert_matches_type(AsyncPage[FileObject], file, path=["response"])
+ assert_matches_type(AsyncCursorPage[FileObject], file, path=["response"])
@parametrize
async def test_method_list_with_all_params(self, async_client: AsyncOpenAI) -> None:
file = await async_client.files.list(
+ after="after",
+ limit=0,
+ order="asc",
purpose="purpose",
)
- assert_matches_type(AsyncPage[FileObject], file, path=["response"])
+ assert_matches_type(AsyncCursorPage[FileObject], file, path=["response"])
@parametrize
async def test_raw_response_list(self, async_client: AsyncOpenAI) -> None:
@@ -350,7 +356,7 @@ async def test_raw_response_list(self, async_client: AsyncOpenAI) -> None:
assert response.is_closed is True
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
file = response.parse()
- assert_matches_type(AsyncPage[FileObject], file, path=["response"])
+ assert_matches_type(AsyncCursorPage[FileObject], file, path=["response"])
@parametrize
async def test_streaming_response_list(self, async_client: AsyncOpenAI) -> None:
@@ -359,7 +365,7 @@ async def test_streaming_response_list(self, async_client: AsyncOpenAI) -> None:
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
file = await response.parse()
- assert_matches_type(AsyncPage[FileObject], file, path=["response"])
+ assert_matches_type(AsyncCursorPage[FileObject], file, path=["response"])
assert cast(Any, response.is_closed) is True