@@ -225,6 +225,10 @@ async def request(
225225 model_request_parameters : ModelRequestParameters ,
226226 ) -> ModelResponse :
227227 check_allow_model_requests ()
228+ model_settings , model_request_parameters = self .prepare_request (
229+ model_settings ,
230+ model_request_parameters ,
231+ )
228232 model_settings = cast (GoogleModelSettings , model_settings or {})
229233 response = await self ._generate_content (messages , False , model_settings , model_request_parameters )
230234 return self ._process_response (response )
@@ -236,6 +240,10 @@ async def count_tokens(
236240 model_request_parameters : ModelRequestParameters ,
237241 ) -> usage .RequestUsage :
238242 check_allow_model_requests ()
243+ model_settings , model_request_parameters = self .prepare_request (
244+ model_settings ,
245+ model_request_parameters ,
246+ )
239247 model_settings = cast (GoogleModelSettings , model_settings or {})
240248 contents , generation_config = await self ._build_content_and_config (
241249 messages , model_settings , model_request_parameters
@@ -291,6 +299,10 @@ async def request_stream(
291299 run_context : RunContext [Any ] | None = None ,
292300 ) -> AsyncIterator [StreamedResponse ]:
293301 check_allow_model_requests ()
302+ model_settings , model_request_parameters = self .prepare_request (
303+ model_settings ,
304+ model_request_parameters ,
305+ )
294306 model_settings = cast (GoogleModelSettings , model_settings or {})
295307 response = await self ._generate_content (messages , True , model_settings , model_request_parameters )
296308 yield await self ._process_streamed_response (response , model_request_parameters ) # type: ignore
0 commit comments