diff --git a/olmocr/bench/scripts/convert_all.sh b/olmocr/bench/scripts/convert_all.sh index 4a3e1cd..586a451 100755 --- a/olmocr/bench/scripts/convert_all.sh +++ b/olmocr/bench/scripts/convert_all.sh @@ -218,14 +218,9 @@ stop_server # python -m olmocr.bench.convert server:name=qwen2_vl_7b:model=Qwen/Qwen2-VL-7B-Instruct:temperature=0.1:prompt_template=full:response_template=plain --repeats 5 --parallel 50 # stop_server -# TODO: qwen2.5 Not working right now in sglang -# qwen25_vl_7b -# create_conda_env "qwen25" "3.11" -# source activate qwen25 -# pip install olmocr -# pip install "sglang[all]>=0.4.3.post2" --find-links https://flashinfer.ai/whl/cu124/torch2.5/flashinfer-python transformers==4.48.3 -# start_server sglang "Qwen/Qwen2.5-VL-7B-Instruct" --chat-template qwen2-vl --mem-fraction-static 0.7 -# python -m olmocr.bench.convert server:name=qwen25_vl_7b:model=Qwen/Qwen2.5-VL-7B-Instruct:temperature=0.1:prompt_template=full:response_template=plain --repeats 5 --parallel 50 +# qwen2.5 works best with vllm for now, in a fresh environment +# start_server vllm "Qwen/Qwen2.5-VL-7B-Instruct" --max-model-len 8192 +# python -m olmocr.bench.convert --dir olmOCR-bench/bench_data server:name=qwen25vl_prompt3:model=Qwen/Qwen2.5-VL-7B-Instruct:temperature=0.1:prompt_template=basic:response_template=plain --parallel 50 # stop_server # TODO: Fix this, I was not able to get it to all install successfully diff --git a/olmocr/bench/tests.py b/olmocr/bench/tests.py index a8065b6..2d18409 100644 --- a/olmocr/bench/tests.py +++ b/olmocr/bench/tests.py @@ -1032,4 +1032,4 @@ def save_tests(tests: List[BasePDFTest], jsonl_file: str) -> None: """ with open(jsonl_file, "w") as file: for test in tests: - file.write(json.dumps(asdict(test)) + "\n") \ No newline at end of file + file.write(json.dumps(asdict(test)) + "\n")