Skip to content

Commit 2e3e625

Browse files
authored
fix tips (#17253)
* fix tips * fix tips
1 parent a9ebda5 commit 2e3e625

File tree

2 files changed

+28
-32
lines changed

2 files changed

+28
-32
lines changed

docs/version3.x/pipeline_usage/PaddleOCR-VL.en.md

Lines changed: 14 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -89,20 +89,16 @@ Currently, PaddleOCR-VL offers four inference methods, with varying levels of su
8989
</tbody>
9090
</table>
9191

92-
TIP:
93-
1. When using NVIDIA GPU for inference, ensure that the Compute Capability (CC) and CUDA version meet the requirements:
94-
95-
- PaddlePaddle: CC ≥ 7.0, CUDA ≥ 11.8
96-
- vLLM: CC ≥ 8.0, CUDA ≥ 12.6
97-
- SGLang: 8.0 ≤ CC < 12.0, CUDA ≥ 12.6
98-
- FastDeploy: 8.0 ≤ CC < 12.0, CUDA ≥ 12.6
99-
- Common GPUs with CC ≥ 8 include RTX 30/40/50 series and A10/A100, etc. For more models, refer to [CUDA GPU Compute Capability](https://developer.nvidia.com/cuda-gpus)
100-
101-
2. vLLM compatibility note: Although vLLM can be launched on NVIDIA GPUs with CC 7.x such as T4/V100, timeout or OOM issues may occur, and its use is not recommended.
102-
103-
3. Currently, PaddleOCR-VL does not support ARM architecture CPUs. More hardware support will be expanded based on actual needs in the future, so stay tuned!
104-
105-
4. vLLM, SGLang, and FastDeploy cannot run natively on Windows or macOS. Please use the Docker images we provide.
92+
> TIP:
93+
> - When using NVIDIA GPU for inference, ensure that the Compute Capability (CC) and CUDA version meet the requirements:
94+
> > - PaddlePaddle: CC ≥ 7.0, CUDA ≥ 11.8
95+
> > - vLLM: CC ≥ 8.0, CUDA ≥ 12.6
96+
> > - SGLang: 8.0 ≤ CC < 12.0, CUDA ≥ 12.6
97+
> > - FastDeploy: 8.0 ≤ CC < 12.0, CUDA ≥ 12.6
98+
> > - Common GPUs with CC ≥ 8 include RTX 30/40/50 series and A10/A100, etc. For more models, refer to [CUDA GPU Compute Capability](https://developer.nvidia.com/cuda-gpus)
99+
> - vLLM compatibility note: Although vLLM can be launched on NVIDIA GPUs with CC 7.x such as T4/V100, timeout or OOM issues may occur, and its use is not recommended.
100+
> - Currently, PaddleOCR-VL does not support ARM architecture CPUs. More hardware support will be expanded based on actual needs in the future, so stay tuned!
101+
> - vLLM, SGLang, and FastDeploy cannot run natively on Windows or macOS. Please use the Docker images we provide.
106102
107103
Since different hardware requires different dependencies, if your hardware meets the requirements in the table above, please refer to the following table for the corresponding tutorial to configure your environment:
108104

@@ -136,7 +132,7 @@ Since different hardware requires different dependencies, if your hardware meets
136132
</tbody>
137133
</table>
138134

139-
> [!TIP]
135+
> TIP:
140136
> For example, if you are using an RTX 50 series GPU that meets the device requirements for both PaddlePaddle and vLLM inference methods, please refer to the [PaddleOCR-VL NVIDIA Blackwell Architecture GPU Environment Configuration Tutorial](./PaddleOCR-VL-NVIDIA-Blackwell.en.md) to complete the environment configuration before using PaddleOCR-VL.
141137
142138
## 1. Environment Preparation
@@ -202,14 +198,14 @@ python -m pip install https://paddle-whl.bj.bcebos.com/nightly/cu126/safetensors
202198
python -m pip install https://xly-devops.cdn.bcebos.com/safetensors-nightly/safetensors-0.6.2.dev0-cp38-abi3-win_amd64.whl
203199
```
204200

205-
> [!IMPORTANT]
201+
> IMPORTANT:
206202
> **Please ensure that you install PaddlePaddle framework version 3.2.1 or above, along with the special version of safetensors.** For macOS users, please use Docker to set up the environment.
207203
208204
## 2. Quick Start
209205

210206
PaddleOCR-VL supports two usage methods: CLI command line and Python API. The CLI command line method is simpler and suitable for quickly verifying functionality, while the Python API method is more flexible and suitable for integration into existing projects.
211207

212-
> [!TIP]
208+
> TIP:
213209
> The methods introduced in this section are primarily for rapid validation. Their inference speed, memory usage, and stability may not meet the requirements of a production environment. **If deployment to a production environment is needed, we strongly recommend using a dedicated inference acceleration framework**. For specific methods, please refer to the next section.
214210
215211
### 2.1 Command Line Usage
@@ -2260,7 +2256,7 @@ foreach ($result as $i => $item) {
22602256

22612257
### 4.4 Pipeline Configuration Adjustment Instructions
22622258

2263-
> [!NOTE]
2259+
> NOTE:
22642260
> If you do not need to adjust pipeline configurations, you can ignore this section.
22652261

22662262
Adjusting the PaddleOCR-VL configuration for service deployment involves only three steps:

docs/version3.x/pipeline_usage/PaddleOCR-VL.md

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -89,16 +89,16 @@ PaddleOCR-VL 是一款先进、高效的文档解析模型,专为文档中的
8989
</tbody>
9090
</table>
9191

92-
> [!TIP]
93-
> 1. 使用英伟达 GPU 推理时需要注意 Compute Capability(简称 CC) 和 CUDA 版本(简称 CUDA)是否满足要求:
94-
> - PaddlePaddle: CC ≥ 7.0, CUDA ≥ 11.8
95-
> - vLLM: CC ≥ 8.0, CUDA ≥ 12.6
96-
> - SGLang: 8.0 ≤ CC < 12.0, CUDA ≥ 12.6
97-
> - FastDeploy: 8.0 ≤ CC < 12.0, CUDA ≥ 12.6
98-
> - CC ≥ 8 的常见显卡包括 RTX 30/40/50 系列及 A10/A100 等,更多型号可查看 [CUDA GPU 计算能力](https://developer.nvidia.cn/cuda-gpus)
99-
> 2. 虽然 vLLM 可在 T4/V100 等 CC 7.x 的 NVIDIA GPU 上启动,但容易出现超时或 OOM,不推荐使用。
100-
> 3. 当前,PaddleOCR-VL 暂不支持 ARM 架构 CPU。后续将根据实际需求扩展更多硬件支持,敬请期待!
101-
> 4. vLLM、SGLang 和 FastDeploy 无法在 Windows 或 macOS 上原生运行,请使用我们提供的 Docker 镜像。
92+
> TIP:
93+
> - 使用英伟达 GPU 推理时需要注意 Compute Capability(简称 CC) 和 CUDA 版本(简称 CUDA)是否满足要求:
94+
> > - PaddlePaddle: CC ≥ 7.0, CUDA ≥ 11.8
95+
> > - vLLM: CC ≥ 8.0, CUDA ≥ 12.6
96+
> > - SGLang: 8.0 ≤ CC < 12.0, CUDA ≥ 12.6
97+
> > - FastDeploy: 8.0 ≤ CC < 12.0, CUDA ≥ 12.6
98+
> > - CC ≥ 8 的常见显卡包括 RTX 30/40/50 系列及 A10/A100 等,更多型号可查看 [CUDA GPU 计算能力](https://developer.nvidia.cn/cuda-gpus)
99+
> - 虽然 vLLM 可在 T4/V100 等 CC 7.x 的 NVIDIA GPU 上启动,但容易出现超时或 OOM,不推荐使用。
100+
> - 当前,PaddleOCR-VL 暂不支持 ARM 架构 CPU。后续将根据实际需求扩展更多硬件支持,敬请期待!
101+
> - vLLM、SGLang 和 FastDeploy 无法在 Windows 或 macOS 上原生运行,请使用我们提供的 Docker 镜像。
102102
103103
由于不同硬件所需的依赖各不相同,如果您的硬件满足上述表格的要求,请参考下表查看对应的教程进行环境配置:
104104

@@ -132,7 +132,7 @@ PaddleOCR-VL 是一款先进、高效的文档解析模型,专为文档中的
132132
</tbody>
133133
</table>
134134

135-
> [!TIP]
135+
> TIP:
136136
> 例如您使用的是 RTX 50 系 GPU,满足 PaddlePaddle 和 vLLM 推理方式的设备要求,请参考 [PaddleOCR-VL NVIDIA Blackwell 架构 GPU 环境配置教程](./PaddleOCR-VL-NVIDIA-Blackwell.md) 完成环境配置后再进行 PaddleOCR-VL 的使用。
137137
138138
## 1. 环境准备
@@ -198,7 +198,7 @@ python -m pip install https://paddle-whl.bj.bcebos.com/nightly/cu126/safetensors
198198
python -m pip install https://xly-devops.cdn.bcebos.com/safetensors-nightly/safetensors-0.6.2.dev0-cp38-abi3-win_amd64.whl
199199
```
200200

201-
> [!IMPORTANT]
201+
> IMPORTANT:
202202
> **请注意安装 3.2.1 及以上版本的飞桨框架,同时安装特殊版本的 safetensors。** 对于 macOS 用户,请使用 Docker 进行环境搭建。
203203
204204
## 2. 快速开始
@@ -207,7 +207,7 @@ python -m pip install https://xly-devops.cdn.bcebos.com/safetensors-nightly/safe
207207

208208
PaddleOCR-VL 支持 CLI 命令行方式和 Python API 两种使用方式,其中 CLI 命令行方式更简单,适合快速验证功能,而 Python API 方式更灵活,适合集成到现有项目中。
209209

210-
> [!TIP]
210+
> TIP:
211211
> 本节所介绍的方法主要用于快速验证,其推理速度、显存占用及稳定性表现未必能满足生产环境的要求。**若需部署至生产环境,我们强烈建议使用专门的推理加速框架** ,具体方法请参考下一节。
212212
213213
### 2.1 命令行方式体验
@@ -2300,7 +2300,7 @@ foreach ($result as $i => $item) {
23002300

23012301
### 4.4 产线配置调整说明
23022302

2303-
> [!NOTE]
2303+
> NOTE:
23042304
> 若您无需调整产线配置,可忽略此小节。
23052305

23062306
调整服务化部署的 PaddleOCR-VL 配置只需以下三步:

0 commit comments

Comments
 (0)