Skip to content

Commit c657f8d

Browse files
authored
[Docs] fix PaddleOCR-VL docs bug (#4702)
1 parent b1dd508 commit c657f8d

File tree

2 files changed

+4
-4
lines changed

2 files changed

+4
-4
lines changed

docs/best_practices/PaddleOCR-VL-0.9B.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ python -m fastdeploy.entrypoints.openai.api_server \
2424
--max-model-len 16384 \
2525
--max-num-batched-tokens 16384 \
2626
--gpu-memory-utilization 0.8 \
27-
--max-num-seqs 128 \
27+
--max-num-seqs 128
2828
```
2929
**Example 2:** Deploying a 16K Context Service on a Single A100 GPU
3030
```shell
@@ -36,7 +36,7 @@ python -m fastdeploy.entrypoints.openai.api_server \
3636
--max-model-len 16384 \
3737
--max-num-batched-tokens 16384 \
3838
--gpu-memory-utilization 0.8 \
39-
--max-num-seqs 256 \
39+
--max-num-seqs 256
4040
```
4141

4242
An example is a set of configurations that can run stably while also delivering relatively good performance. If you have further requirements for precision or performance, please continue reading the content below.

docs/zh/best_practices/PaddleOCR-VL-0.9B.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ python -m fastdeploy.entrypoints.openai.api_server \
2424
--max-model-len 16384 \
2525
--max-num-batched-tokens 16384 \
2626
--gpu-memory-utilization 0.8 \
27-
--max-num-seqs 128 \
27+
--max-num-seqs 128
2828
```
2929

3030
**示例2:** A100上单卡部署16K上下文的服务
@@ -37,7 +37,7 @@ python -m fastdeploy.entrypoints.openai.api_server \
3737
--max-model-len 16384 \
3838
--max-num-batched-tokens 16384 \
3939
--gpu-memory-utilization 0.8 \
40-
--max-num-seqs 256 \
40+
--max-num-seqs 256
4141
```
4242

4343
示例是可以稳定运行的一组配置,同时也能得到比较好的性能。

0 commit comments

Comments
 (0)