|
|
@@ -4,6 +4,7 @@ import time
|
|
|
|
|
|
from loguru import logger
|
|
|
|
|
|
+from .custom_logits_processors import enable_custom_logits_processors
|
|
|
from .model_output_to_middle_json import result_to_middle_json
|
|
|
from ...data.data_reader_writer import DataWriter
|
|
|
from mineru.utils.pdf_image_tools import load_images_from_pdf
|
|
|
@@ -43,15 +44,8 @@ class ModelSingleton:
|
|
|
batch_size = 0
|
|
|
if backend in ['transformers', 'vllm-engine', "vllm-async-engine"] and not model_path:
|
|
|
model_path = auto_download_and_get_model_root_path("/","vlm")
|
|
|
- import torch
|
|
|
- compute_capability = 0.0
|
|
|
- custom_logits_processors = False
|
|
|
- if torch.cuda.is_available():
|
|
|
- major, minor = torch.cuda.get_device_capability()
|
|
|
- compute_capability = float(major) + (float(minor) / 10.0)
|
|
|
- logger.info(f"compute_capability: {compute_capability}")
|
|
|
- if compute_capability >= 8.0:
|
|
|
- custom_logits_processors = True
|
|
|
+
|
|
|
+ custom_logits_processors = enable_custom_logits_processors()
|
|
|
|
|
|
if backend == "transformers":
|
|
|
try:
|