@@ -660,7 +660,7 @@ async def _process_streamed_response(
660660
661661 # When using Azure OpenAI and a content filter is enabled, the first chunk will contain a `''` model name,
662662 # so we set it from a later chunk in `OpenAIChatStreamedResponse`.
663- model_name = first_chunk .model or self ._model_name
663+ model_name = first_chunk .model or self .model_name
664664
665665 return self ._streamed_response_cls (
666666 model_request_parameters = model_request_parameters ,
@@ -681,7 +681,7 @@ def _streamed_response_cls(self) -> type[OpenAIStreamedResponse]:
681681 return OpenAIStreamedResponse
682682
683683 def _map_usage (self , response : chat .ChatCompletion ) -> usage .RequestUsage :
684- return _map_usage (response , self ._provider .name , self ._provider .base_url , self ._model_name )
684+ return _map_usage (response , self ._provider .name , self ._provider .base_url , self .model_name )
685685
686686 def _get_tools (self , model_request_parameters : ModelRequestParameters ) -> list [chat .ChatCompletionToolParam ]:
687687 return [self ._map_tool_definition (r ) for r in model_request_parameters .tool_defs .values ()]
@@ -1216,7 +1216,7 @@ def _process_response( # noqa: C901
12161216
12171217 return ModelResponse (
12181218 parts = items ,
1219- usage = _map_usage (response , self ._provider .name , self ._provider .base_url , self ._model_name ),
1219+ usage = _map_usage (response , self ._provider .name , self ._provider .base_url , self .model_name ),
12201220 model_name = response .model ,
12211221 provider_response_id = response .id ,
12221222 timestamp = timestamp ,
@@ -1341,7 +1341,7 @@ async def _responses_create(
13411341 extra_headers .setdefault ('User-Agent' , get_user_agent ())
13421342 return await self .client .responses .create (
13431343 input = openai_messages ,
1344- model = self ._model_name ,
1344+ model = self .model_name ,
13451345 instructions = instructions ,
13461346 parallel_tool_calls = model_settings .get ('parallel_tool_calls' , OMIT ),
13471347 tools = tools or OMIT ,
@@ -1943,7 +1943,7 @@ def _map_provider_details(self, chunk: ChatCompletionChunk) -> dict[str, Any] |
19431943 return _map_provider_details (chunk .choices [0 ])
19441944
19451945 def _map_usage (self , response : ChatCompletionChunk ) -> usage .RequestUsage :
1946- return _map_usage (response , self ._provider_name , self ._provider_url , self ._model_name )
1946+ return _map_usage (response , self ._provider_name , self ._provider_url , self .model_name )
19471947
19481948 def _map_finish_reason (
19491949 self , key : Literal ['stop' , 'length' , 'tool_calls' , 'content_filter' , 'function_call' ]
@@ -2265,7 +2265,7 @@ async def _get_event_iterator(self) -> AsyncIterator[ModelResponseStreamEvent]:
22652265 )
22662266
22672267 def _map_usage (self , response : responses .Response ) -> usage .RequestUsage :
2268- return _map_usage (response , self ._provider_name , self ._provider_url , self ._model_name )
2268+ return _map_usage (response , self ._provider_name , self ._provider_url , self .model_name )
22692269
22702270 @property
22712271 def model_name (self ) -> OpenAIModelName :
0 commit comments