Update README.md
Browse files
README.md
CHANGED
|
@@ -168,7 +168,7 @@ We suggest using `vllm>=0.8.5` and enabling long context in VLLM to serve DeepSW
|
|
| 168 |
```bash
|
| 169 |
export MAX_CONTEXT_LEN=65536
|
| 170 |
export TENSOR_PARALLEL_SIZE=8
|
| 171 |
-
VLLM_ALLOW_LONG_MAX_MODEL_LEN=1 vllm serve agentica-org/DeepSWE-Preview --tensor-parallel-size $TENSOR_PARALLEL_SIZE --max-model-len $MAX_CONTEXT_LEN --hf-overrides '{"max_position_embeddings": $MAX_CONTEXT_LEN}' --enable_prefix_caching
|
| 172 |
```
|
| 173 |
|
| 174 |
## License
|
|
|
|
| 168 |
```bash
|
| 169 |
export MAX_CONTEXT_LEN=65536
|
| 170 |
export TENSOR_PARALLEL_SIZE=8
|
| 171 |
+
VLLM_ALLOW_LONG_MAX_MODEL_LEN=1 vllm serve agentica-org/DeepSWE-Preview --tensor-parallel-size $TENSOR_PARALLEL_SIZE --max-model-len $MAX_CONTEXT_LEN --hf-overrides '{\"max_position_embeddings\": $MAX_CONTEXT_LEN}' --enable_prefix_caching
|
| 172 |
```
|
| 173 |
|
| 174 |
## License
|