mirror of
https://github.com/allenai/olmocr.git
synced 2025-11-04 03:56:16 +00:00
Bf16 only
This commit is contained in:
parent
44dd966850
commit
bde6f2955e
@ -147,11 +147,6 @@ class TrainingConfig:
|
|||||||
gradient_checkpointing: bool = True
|
gradient_checkpointing: bool = True
|
||||||
gradient_checkpointing_kwargs: Dict[str, Any] = field(default_factory=lambda: {"use_reentrant": False})
|
gradient_checkpointing_kwargs: Dict[str, Any] = field(default_factory=lambda: {"use_reentrant": False})
|
||||||
|
|
||||||
# Mixed precision
|
|
||||||
fp16: bool = False
|
|
||||||
bf16: bool = True
|
|
||||||
tf32: bool = True # Enable TF32 on Ampere GPUs
|
|
||||||
|
|
||||||
# Evaluation and checkpointing
|
# Evaluation and checkpointing
|
||||||
evaluation_strategy: str = "steps"
|
evaluation_strategy: str = "steps"
|
||||||
eval_steps: int = 500
|
eval_steps: int = 500
|
||||||
|
|||||||
@ -178,9 +178,7 @@ def main():
|
|||||||
adam_epsilon=config.training.adam_epsilon,
|
adam_epsilon=config.training.adam_epsilon,
|
||||||
weight_decay=config.training.weight_decay,
|
weight_decay=config.training.weight_decay,
|
||||||
max_grad_norm=config.training.max_grad_norm,
|
max_grad_norm=config.training.max_grad_norm,
|
||||||
fp16=config.training.fp16,
|
bf16=True, # We're sticking with this known good reduced precision option
|
||||||
bf16=config.training.bf16,
|
|
||||||
tf32=config.training.tf32,
|
|
||||||
eval_strategy=config.training.evaluation_strategy,
|
eval_strategy=config.training.evaluation_strategy,
|
||||||
eval_steps=config.training.eval_steps,
|
eval_steps=config.training.eval_steps,
|
||||||
save_strategy=config.training.save_strategy,
|
save_strategy=config.training.save_strategy,
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user