| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354 |
- """
- bench
- """
- import os
- import shutil
- import json
- from lib import calculate_score
- import pytest
- from conf import conf
- code_path = os.environ.get('GITHUB_WORKSPACE')
- pdf_dev_path = conf.conf["pdf_dev_path"]
- pdf_res_path = conf.conf["pdf_res_path"]
- class TestBench():
- """
- test bench
- """
- def test_ci_ben(self):
- """
- ci benchmark
- """
- fr = open(os.path.join(pdf_dev_path, "result.json"), "r", encoding="utf-8")
- lines = fr.readlines()
- last_line = lines[-1].strip()
- last_score = json.loads(last_line)
- last_simscore = last_score["average_sim_score"]
- last_editdistance = last_score["average_edit_distance"]
- last_bleu = last_score["average_bleu_score"]
- os.system(f"python tests/test_cli/lib/pre_clean.py --tool_name mineru --download_dir {pdf_dev_path}")
- now_score = get_score()
- print ("now_score:", now_score)
- if not os.path.exists(os.path.join(pdf_dev_path, "ci")):
- os.makedirs(os.path.join(pdf_dev_path, "ci"), exist_ok=True)
- fw = open(os.path.join(pdf_dev_path, "ci", "result.json"), "w+", encoding="utf-8")
- fw.write(json.dumps(now_score) + "\n")
- now_simscore = now_score["average_sim_score"]
- now_editdistance = now_score["average_edit_distance"]
- now_bleu = now_score["average_bleu_score"]
- assert last_simscore <= now_simscore
- assert last_editdistance <= now_editdistance
- assert last_bleu <= now_bleu
- def get_score():
- """
- get score
- """
- score = calculate_score.Scoring(os.path.join(pdf_dev_path, "result.json"))
- score.calculate_similarity_total("mineru", pdf_dev_path)
- res = score.summary_scores()
- return res
|