From e509a0adc556b1af8bf3af6140fd67ceb4a53c70 Mon Sep 17 00:00:00 2001 From: root Date: Wed, 30 Jul 2025 15:04:15 +0000 Subject: [PATCH] Add vLLM support for gemma-3-1b-it --- examples/offline_inference_tt.py | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/examples/offline_inference_tt.py b/examples/offline_inference_tt.py index a4fb4fa6b8b6..91ca8b0dbe42 100644 --- a/examples/offline_inference_tt.py +++ b/examples/offline_inference_tt.py @@ -55,7 +55,10 @@ def register_tt_models(): ModelRegistry.register_model( "TTMistralForCausalLM", "models.tt_transformers.tt.generator_vllm:MistralForCausalLM") - + ModelRegistry.register_model( + "TTGemma3ForCausalLM", + "models.tt_transformers.tt.generator_vllm:Gemma3ForCausalLM" + ) register_tt_models() # Import and register models from tt-metal @@ -121,6 +124,7 @@ def check_tt_model_supported(model): "deepseek-ai/DeepSeek-R1-Distill-Llama-70B", "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B", "mistralai/Mistral-7B-Instruct-v0.3", + "google/gemma-3-1b-it", ] assert model in supported_models, f"Invalid model: {model}"