Bläddra i källkod

Merge pull request #3506 from myhloli/dev

Dev
Xiaomeng Zhao 2 månader sedan
förälder
incheckning
9af1824328

+ 10 - 12
docs/en/reference/output_files.md

@@ -698,18 +698,16 @@ Example: discarded blocks output
 
 The above files constitute MinerU's complete output results. Users can choose appropriate files for subsequent processing based on their needs:
 
-- **Model outputs**: 
-
-  * Use raw outputs (model.json, model_output.txt)
+- **Model outputs** (Use raw outputs):  
+    * model.json
   
-- **Debugging and verification**:
-
-  * Use visualization files (layout.pdf, spans.pdf) 
+- **Debugging and verification** (Use visualization files):
+    * layout.pdf
+    * spans.pdf 
   
-- **Content extraction**: 
-
-  * Use simplified files (*.md, content_list.json)
+- **Content extraction**: (Use simplified files):
+    * *.md
+    * content_list.json
   
-- **Secondary development**: 
-
-  * Use structured files (middle.json)
+- **Secondary development**: (Use structured files):
+    * middle.json

+ 14 - 20
docs/en/usage/cli_tools.md

@@ -66,33 +66,27 @@ Some parameters of MinerU command line tools have equivalent environment variabl
 Here are the environment variables and their descriptions:
 
 - `MINERU_DEVICE_MODE`:
-
-  * Used to specify inference device
-  * supports device types like `cpu/cuda/cuda:0/npu/mps`
-  * only effective for `pipeline` backend.
+    * Used to specify inference device
+    * supports device types like `cpu/cuda/cuda:0/npu/mps`
+    * only effective for `pipeline` backend.
   
 - `MINERU_VIRTUAL_VRAM_SIZE`: 
-
-  * Used to specify maximum GPU VRAM usage per process (GB)
-  * only effective for `pipeline` backend.
+    * Used to specify maximum GPU VRAM usage per process (GB)
+    * only effective for `pipeline` backend.
   
 - `MINERU_MODEL_SOURCE`: 
-
-  * Used to specify model source
-  * supports `huggingface/modelscope/local`
-  * defaults to `huggingface`, can be switched to `modelscope` or local models through environment variables.
+    * Used to specify model source
+    * supports `huggingface/modelscope/local`
+    * defaults to `huggingface`, can be switched to `modelscope` or local models through environment variables.
   
 - `MINERU_TOOLS_CONFIG_JSON`: 
-
-  * Used to specify configuration file path
-  * defaults to `mineru.json` in user directory, can specify other configuration file paths through environment variables.
+    * Used to specify configuration file path
+    * defaults to `mineru.json` in user directory, can specify other configuration file paths through environment variables.
   
 - `MINERU_FORMULA_ENABLE`:
-
-  * Used to enable formula parsing
-  * defaults to `true`, can be set to `false` through environment variables to disable formula parsing.
+    * Used to enable formula parsing
+    * defaults to `true`, can be set to `false` through environment variables to disable formula parsing.
   
 - `MINERU_TABLE_ENABLE`: 
-
-  * Used to enable table parsing
-  * defaults to `true`, can be set to `false` through environment variables to disable table parsing.
+    * Used to enable table parsing
+    * defaults to `true`, can be set to `false` through environment variables to disable table parsing.

+ 8 - 12
docs/en/usage/quick_usage.md

@@ -78,19 +78,15 @@ MinerU is now ready to use out of the box, but also supports extending functiona
 Here are some available configuration options:  
 
 - `latex-delimiter-config`: 
-
-  * Used to configure LaTeX formula delimiters
-  * Defaults to `$` symbol, can be modified to other symbols or strings as needed.
+    * Used to configure LaTeX formula delimiters
+    * Defaults to `$` symbol, can be modified to other symbols or strings as needed.
   
 - `llm-aided-config`:
-
-  * Used to configure parameters for LLM-assisted title hierarchy
-  * Compatible with all LLM models supporting `openai protocol`, defaults to using Alibaba Cloud Bailian's `qwen2.5-32b-instruct` model. 
-  * You need to configure your own API key and set `enable` to `true` to enable this feature.
+    * Used to configure parameters for LLM-assisted title hierarchy
+    * Compatible with all LLM models supporting `openai protocol`, defaults to using Alibaba Cloud Bailian's `qwen2.5-32b-instruct` model. 
+    * You need to configure your own API key and set `enable` to `true` to enable this feature.
   
 - `models-dir`: 
-
-  * Used to specify local model storage directory
-  * Please specify model directories for `pipeline` and `vlm` backends separately.
-  * After specifying the directory, you can use local models by configuring the environment variable `export MINERU_MODEL_SOURCE=local`.
-
+    * Used to specify local model storage directory
+    * Please specify model directories for `pipeline` and `vlm` backends separately.
+    * After specifying the directory, you can use local models by configuring the environment variable `export MINERU_MODEL_SOURCE=local`.

+ 10 - 12
docs/zh/reference/output_files.md

@@ -808,18 +808,16 @@ vlm 后端的 content_list.json 文件结构与 pipeline 后端类似,伴随
 
 以上文件为 MinerU 的完整输出结果,用户可根据需要选择合适的文件进行后续处理:
 
-- **模型输出**:
-
-  * 使用原始输出(model.json、model_output.txt)
+- **模型输出**(使用原始输出):
+    * model.json
   
-- **调试和验证**
-
-  * 使用可视化文件(layout.pdf、spans.pdf 
+- **调试和验证**(使用可视化文件):
+    * layout.pdf
+    * spans.pdf 
   
-- **内容提取**
-
-  * 使用简化文件(*.md、content_list.json)
+- **内容提取**(使用简化文件):
+    * *.md
+    * content_list.json
   
-- **二次开发**:
-
-  * 使用结构化文件(middle.json)
+- **二次开发**(使用结构化文件):
+    * middle.json

+ 14 - 20
docs/zh/usage/cli_tools.md

@@ -61,33 +61,27 @@ MinerU命令行工具的某些参数存在相同功能的环境变量配置,
 以下是常用的环境变量及其说明: 
 
 - `MINERU_DEVICE_MODE`:
-
-  * 用于指定推理设备
-  * 支持`cpu/cuda/cuda:0/npu/mps`等设备类型
-  * 仅对`pipeline`后端生效。
+    * 用于指定推理设备
+    * 支持`cpu/cuda/cuda:0/npu/mps`等设备类型
+    * 仅对`pipeline`后端生效。
   
 - `MINERU_VIRTUAL_VRAM_SIZE`:
-
-  * 用于指定单进程最大 GPU 显存占用(GB)
-  * 仅对`pipeline`后端生效。
+    * 用于指定单进程最大 GPU 显存占用(GB)
+    * 仅对`pipeline`后端生效。
   
 - `MINERU_MODEL_SOURCE`:
-
-  * 用于指定模型来源
-  * 支持`huggingface/modelscope/local`
-  * 默认为`huggingface`可通过环境变量切换为`modelscope`或使用本地模型。
+    * 用于指定模型来源
+    * 支持`huggingface/modelscope/local`
+    * 默认为`huggingface`可通过环境变量切换为`modelscope`或使用本地模型。
   
 - `MINERU_TOOLS_CONFIG_JSON`:
-
-  * 用于指定配置文件路径
-  * 默认为用户目录下的`mineru.json`,可通过环境变量指定其他配置文件路径。
+    * 用于指定配置文件路径
+    * 默认为用户目录下的`mineru.json`,可通过环境变量指定其他配置文件路径。
   
 - `MINERU_FORMULA_ENABLE`:
-
-  * 用于启用公式解析
-  * 默认为`true`,可通过环境变量设置为`false`来禁用公式解析。
+    * 用于启用公式解析
+    * 默认为`true`,可通过环境变量设置为`false`来禁用公式解析。
   
 - `MINERU_TABLE_ENABLE`:
-
-  * 用于启用表格解析
-  * 默认为`true`,可通过环境变量设置为`false`来禁用表格解析。
+    * 用于启用表格解析
+    * 默认为`true`,可通过环境变量设置为`false`来禁用表格解析。

+ 7 - 10
docs/zh/usage/quick_usage.md

@@ -77,17 +77,14 @@ MinerU 现已实现开箱即用,但也支持通过配置文件扩展功能。
 以下是一些可用的配置选项: 
 
 - `latex-delimiter-config`:
-
-  * 用于配置 LaTeX 公式的分隔符
-  * 默认为`$`符号,可根据需要修改为其他符号或字符串。
+    * 用于配置 LaTeX 公式的分隔符
+    * 默认为`$`符号,可根据需要修改为其他符号或字符串。
   
 - `llm-aided-config`:
-
-  * 用于配置 LLM 辅助标题分级的相关参数,兼容所有支持`openai协议`的 LLM 模型
-  * 默认使用`阿里云百炼`的`qwen2.5-32b-instruct`模型
-  * 您需要自行配置 API 密钥并将`enable`设置为`true`来启用此功能。
+    * 用于配置 LLM 辅助标题分级的相关参数,兼容所有支持`openai协议`的 LLM 模型
+    * 默认使用`阿里云百炼`的`qwen2.5-32b-instruct`模型
+    * 您需要自行配置 API 密钥并将`enable`设置为`true`来启用此功能。
   
 - `models-dir`:
-
-  * 用于指定本地模型存储目录,请为`pipeline`和`vlm`后端分别指定模型目录,
-  * 指定目录后您可通过配置环境变量`export MINERU_MODEL_SOURCE=local`来使用本地模型。
+    * 用于指定本地模型存储目录,请为`pipeline`和`vlm`后端分别指定模型目录,
+    * 指定目录后您可通过配置环境变量`export MINERU_MODEL_SOURCE=local`来使用本地模型。

+ 1 - 1
mkdocs.yml

@@ -105,7 +105,6 @@ plugins:
       reconfigure_material: true
       docs_structure: folder
       fallback_to_default: true
-      reconfigure_material: true
       reconfigure_search: true
       languages:
         - locale: en
@@ -135,6 +134,7 @@ markdown_extensions:
   - admonition
   - pymdownx.details
   - attr_list
+  - def_list
   - gfm_admonition
   - pymdownx.highlight:
       use_pygments: true