Skip to content

Commit 5eb7778

Browse files
authored
Replace direct _model_name access with model_name property in OpenAI models and streamed responses (#3564)
1 parent f542189 commit 5eb7778

File tree

1 file changed

+6
-6
lines changed

1 file changed

+6
-6
lines changed

pydantic_ai_slim/pydantic_ai/models/openai.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -660,7 +660,7 @@ async def _process_streamed_response(
660660

661661
# When using Azure OpenAI and a content filter is enabled, the first chunk will contain a `''` model name,
662662
# so we set it from a later chunk in `OpenAIChatStreamedResponse`.
663-
model_name = first_chunk.model or self._model_name
663+
model_name = first_chunk.model or self.model_name
664664

665665
return self._streamed_response_cls(
666666
model_request_parameters=model_request_parameters,
@@ -681,7 +681,7 @@ def _streamed_response_cls(self) -> type[OpenAIStreamedResponse]:
681681
return OpenAIStreamedResponse
682682

683683
def _map_usage(self, response: chat.ChatCompletion) -> usage.RequestUsage:
684-
return _map_usage(response, self._provider.name, self._provider.base_url, self._model_name)
684+
return _map_usage(response, self._provider.name, self._provider.base_url, self.model_name)
685685

686686
def _get_tools(self, model_request_parameters: ModelRequestParameters) -> list[chat.ChatCompletionToolParam]:
687687
return [self._map_tool_definition(r) for r in model_request_parameters.tool_defs.values()]
@@ -1216,7 +1216,7 @@ def _process_response( # noqa: C901
12161216

12171217
return ModelResponse(
12181218
parts=items,
1219-
usage=_map_usage(response, self._provider.name, self._provider.base_url, self._model_name),
1219+
usage=_map_usage(response, self._provider.name, self._provider.base_url, self.model_name),
12201220
model_name=response.model,
12211221
provider_response_id=response.id,
12221222
timestamp=timestamp,
@@ -1341,7 +1341,7 @@ async def _responses_create(
13411341
extra_headers.setdefault('User-Agent', get_user_agent())
13421342
return await self.client.responses.create(
13431343
input=openai_messages,
1344-
model=self._model_name,
1344+
model=self.model_name,
13451345
instructions=instructions,
13461346
parallel_tool_calls=model_settings.get('parallel_tool_calls', OMIT),
13471347
tools=tools or OMIT,
@@ -1943,7 +1943,7 @@ def _map_provider_details(self, chunk: ChatCompletionChunk) -> dict[str, Any] |
19431943
return _map_provider_details(chunk.choices[0])
19441944

19451945
def _map_usage(self, response: ChatCompletionChunk) -> usage.RequestUsage:
1946-
return _map_usage(response, self._provider_name, self._provider_url, self._model_name)
1946+
return _map_usage(response, self._provider_name, self._provider_url, self.model_name)
19471947

19481948
def _map_finish_reason(
19491949
self, key: Literal['stop', 'length', 'tool_calls', 'content_filter', 'function_call']
@@ -2265,7 +2265,7 @@ async def _get_event_iterator(self) -> AsyncIterator[ModelResponseStreamEvent]:
22652265
)
22662266

22672267
def _map_usage(self, response: responses.Response) -> usage.RequestUsage:
2268-
return _map_usage(response, self._provider_name, self._provider_url, self._model_name)
2268+
return _map_usage(response, self._provider_name, self._provider_url, self.model_name)
22692269

22702270
@property
22712271
def model_name(self) -> OpenAIModelName:

0 commit comments

Comments
 (0)