mirror of
https://github.com/rasbt/LLMs-from-scratch.git
synced 2025-10-27 07:49:25 +00:00
use torch no grad for loss (#119)
This commit is contained in:
parent
98f1e97452
commit
ef2de4718e
@ -1135,8 +1135,10 @@
|
|||||||
"\n",
|
"\n",
|
||||||
"\n",
|
"\n",
|
||||||
"torch.manual_seed(123) # For reproducibility due to the shuffling in the data loader\n",
|
"torch.manual_seed(123) # For reproducibility due to the shuffling in the data loader\n",
|
||||||
"train_loss = calc_loss_loader(train_loader, model, device)\n",
|
"\n",
|
||||||
"val_loss = calc_loss_loader(val_loader, model, device)\n",
|
"with torch.no_grad(): # Disable gradient tracking for efficiency because we are not training, yet\n",
|
||||||
|
" train_loss = calc_loss_loader(train_loader, model, device)\n",
|
||||||
|
" val_loss = calc_loss_loader(val_loader, model, device)\n",
|
||||||
"\n",
|
"\n",
|
||||||
"print(\"Training loss:\", train_loss)\n",
|
"print(\"Training loss:\", train_loss)\n",
|
||||||
"print(\"Validation loss:\", val_loss)"
|
"print(\"Validation loss:\", val_loss)"
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user