diff --git a/olmocr/train/configs/example_config.yaml b/olmocr/train/configs/example_config.yaml index 1508de0..2c1aa78 100644 --- a/olmocr/train/configs/example_config.yaml +++ b/olmocr/train/configs/example_config.yaml @@ -8,20 +8,20 @@ run_name: qwen2.5-vl-7b-finetune model: name: Qwen/Qwen2.5-VL-7B-Instruct trust_remote_code: true - torch_dtype: auto + torch_dtype: bfloat16 use_flash_attention: true attn_implementation: flash_attention_2 # LoRA settings (disabled by default) use_lora: false - lora_rank: 8 - lora_alpha: 32 - lora_dropout: 0.1 - lora_target_modules: - - q_proj - - v_proj - - k_proj - - o_proj + # lora_rank: 8 + # lora_alpha: 32 + # lora_dropout: 0.1 + # lora_target_modules: + # - q_proj + # - v_proj + # - k_proj + # - o_proj # Dataset configuration dataset: