Skip to content

Commit 6e40c3d

Browse files
authored
Update README.md
1 parent c59d7fa commit 6e40c3d

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

vllm/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ llm-scaler-vllm is an extended and optimized version of vLLM, specifically adapt
2020
2.4 [Multi-Modal Model Support](#24-multi-modal-model-support)
2121
2.5 [Omni Model Support](#25-omni-model-support)
2222
2.6 [Data Parallelism (DP)](#26-data-parallelism-dp)
23-
2.7 [Maximum Context Length Support](#27-maximum-context-length-support)
23+
2.7 [Finding maximum Context Length](#27-finding-maximum-context-length)
2424
3. [Supported Models](#3-supported-models)
2525
4. [Troubleshooting](#4-troubleshooting)
2626
5. [Performance tuning](#5-performance-tuning)
@@ -2281,7 +2281,7 @@ To enable data parallelism, add:
22812281

22822282
---
22832283

2284-
### 2.7 Maximum Context Length Support
2284+
### 2.7 Finding maximum Context Length
22852285
When using the `V1` engine, the system automatically logs the maximum supported context length during startup based on the available GPU memory and KV cache configuration.
22862286

22872287
#### Example: Successful Startup

0 commit comments

Comments
 (0)