Sfoglia il codice sorgente

feat(更新GLM-OCR本地守护进程配置): 修改端口和模型路径,增强本地服务兼容性

zhch158_admin 1 settimana fa
parent
commit
d5b79e5f52
1 ha cambiato i file con 10 aggiunte e 5 eliminazioni
  1. 10 5
      ocr_tools/daemons/glmocr_local_daemon.sh

+ 10 - 5
ocr_tools/daemons/glmocr_local_daemon.sh

@@ -5,7 +5,12 @@
 # 模型下载地址: https://huggingface.co/ggml-org/GLM-OCR-GGUF
 # 模型下载地址: https://huggingface.co/PaddlePaddle/PaddleOCR-VL-1.5-GGUF
 
-# curl -X POST http://localhost:8080/v1/chat/completions -d @payload.json
+# unset https_proxy http_proxy HF_ENDPOINT
+# llama-server -hf ggml-org/GLM-OCR-GGUF:Q8_0
+# mv ~/Library/Caches/llama.cpp/ggml-org_GLM-OCR-GGUF_GLM-OCR-Q8_0.gguf  ~/models/glmocr
+# mv ~/Library/Caches/llama.cpp/ggml-org_GLM-OCR-GGUF_mmproj-GLM-OCR-Q8_0.gguf  ~/models/glmocr
+
+# curl -X POST http://localhost:8101/v1/chat/completions -d @payload.json
 
 LOGDIR="$HOME/workspace/logs"
 mkdir -p $LOGDIR
@@ -14,12 +19,12 @@ LOGFILE="$LOGDIR/glmocr_llamaserver.log"
 
 # 配置参数
 CONDA_ENV="mineru2"
-PORT="8080"
+PORT="8101"
 HOST="0.0.0.0"
 
 # 本地 GGUF 模型路径
-MODEL_PATH="$HOME/Library/Caches/llama.cpp/ggml-org_GLM-OCR-GGUF_GLM-OCR-Q8_0.gguf"
-MMPROJ_PATH="$HOME/Library/Caches/llama.cpp/ggml-org_GLM-OCR-GGUF_mmproj-GLM-OCR-Q8_0.gguf"
+MODEL_PATH="$HOME/models/glmocr/ggml-org_GLM-OCR-GGUF_GLM-OCR-Q8_0.gguf"
+MMPROJ_PATH="$HOME/models/glmocr/ggml-org_GLM-OCR-GGUF_mmproj-GLM-OCR-Q8_0.gguf"
 
 # llama-server 参数
 CONTEXT_SIZE="16384"         # 上下文长度(需 >= max_tokens,推荐 8192-16384)
@@ -92,7 +97,7 @@ start() {
         --mmproj "$MMPROJ_PATH" \
         --host $HOST \
         --port $PORT \
-        --media-path /Users/zhch158/workspace \
+        --media-path $HOME/workspace \
         -c $CONTEXT_SIZE \
         -ngl $GPU_LAYERS \
         -t $THREADS \