11# SPDX-License-Identifier: Apache-2.0
22# SPDX-FileCopyrightText: Copyright contributors to the vLLM project
33
4+ import pytest
5+
46from vllm import SamplingParams
57from vllm .config .load import LoadConfig
68from vllm .model_executor .model_loader import get_model_loader
79
810load_format = "runai_streamer"
911test_model = "openai-community/gpt2"
12+ # TODO(amacaskill): Replace with a GKE owned GCS bucket.
13+ test_gcs_model = "gs://vertex-model-garden-public-us/codegemma/codegemma-2b/"
1014
1115prompts = [
1216 "Hello, my name is" ,
@@ -32,3 +36,16 @@ def test_runai_model_loader_download_files(vllm_runner):
3236 with vllm_runner (test_model , load_format = load_format ) as llm :
3337 deserialized_outputs = llm .generate (prompts , sampling_params )
3438 assert deserialized_outputs
39+
40+
41+ def test_runai_model_loader_download_files_gcs (
42+ vllm_runner , monkeypatch : pytest .MonkeyPatch
43+ ):
44+ monkeypatch .setenv ("GOOGLE_CLOUD_PROJECT" , "fake-project" )
45+ monkeypatch .setenv ("RUNAI_STREAMER_GCS_USE_ANONYMOUS_CREDENTIALS" , "true" )
46+ monkeypatch .setenv (
47+ "CLOUD_STORAGE_EMULATOR_ENDPOINT" , "https://storage.googleapis.com"
48+ )
49+ with vllm_runner (test_gcs_model , load_format = load_format ) as llm :
50+ deserialized_outputs = llm .generate (prompts , sampling_params )
51+ assert deserialized_outputs
0 commit comments