diff --git a/README.md b/README.md index 91ae6d6..0022a10 100644 --- a/README.md +++ b/README.md @@ -54,7 +54,9 @@ Default: 9192 ### --model Specifies the model to load. This will be downloaded automatically if it does not exist. +\ Default: Qwen2.5-VL-7B-Instruct +\ Choices: Qwen2.5-VL-3B-Instruct, Qwen2.5-VL-7B-Instruct, Qwen2.5-VL-72B-Instruct ### --resume @@ -128,8 +130,7 @@ curl http://localhost:9192/health Environment variables in docker-compose.yml: - `NVIDIA_VISIBLE_DEVICES`: GPU device selection -- `MODEL_DIR`: Model directory path -- `PORT`: API port (default: 9192) +- `QWEN_MODEL`: Select the Qwen 2.5 VL model to load ## Integration with OpenWebUI @@ -228,22 +229,3 @@ This project is licensed under the MIT License - see the LICENSE file for detail ## Support For issues and feature requests, please use the GitHub issue tracker. -``` - -This README provides: -1. Clear installation instructions -2. API documentation -3. Configuration options -4. System requirements -5. Usage examples -6. Development guidelines -7. Monitoring information -8. Error handling details -9. Contributing guidelines - -You may want to customize: -- Repository URLs -- License information -- Specific system requirements based on your deployment -- Additional configuration options -- Any specific deployment instructions for your environment diff --git a/docker-compose.yml b/docker-compose.yml index f90aab1..acd7019 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -14,7 +14,7 @@ services: count: all capabilities: [gpu] environment: - - NVIDIA_VISIBLE_DEVICES=all + - NVIDIA_VISIBLE_DEVICES=${NVIDIA_VISIBLE_DEVICES:-"all"} - QWEN_MODEL: ${QWEN_MODEL:-"Qwen2.5-VL-7B-Instruct"} restart: unless-stopped healthcheck: