Signed-off-by: Roger Wang <hey@rogerw.io>
@@ -1030,7 +1030,8 @@ We highly recommend using vLLM for deployment and inference. All of our evaluati
# Launch vLLM model server
vllm serve rednote-hilab/dots.ocr --trust-remote-code --async-scheduling
-# vllm API Demo
+# vLLM API Demo
+# See demo/demo_vllm.py for details on parameter and prompt settings that help achieve the best output quality.
python3 ./demo/demo_vllm.py --prompt_mode prompt_layout_all_en
```