mirror of
https://github.com/rasbt/LLMs-from-scratch.git
synced 2025-12-03 10:30:50 +00:00
Ch06 classifier function asserts (#703)
This commit is contained in:
parent
f5bc863752
commit
4014bdd520
@ -2353,7 +2353,17 @@
|
||||
"\n",
|
||||
" # Truncate sequences if they too long\n",
|
||||
" input_ids = input_ids[:min(max_length, supported_context_length)]\n",
|
||||
"\n",
|
||||
" assert max_length is not None, (\n",
|
||||
" \"max_length must be specified. If you want to use the full model context, \"\n",
|
||||
" \"pass max_length=model.pos_emb.weight.shape[0].\"\n",
|
||||
" )\n",
|
||||
" assert max_length <= supported_context_length, (\n",
|
||||
" f\"max_length ({max_length}) exceeds model's supported context length ({supported_context_length}).\"\n",
|
||||
" ) \n",
|
||||
" # Alternatively, a more robust version is the following one, which handles the max_length=None case better\n",
|
||||
" # max_len = min(max_length,supported_context_length) if max_length else supported_context_length\n",
|
||||
" # input_ids = input_ids[:max_len]\n",
|
||||
" \n",
|
||||
" # Pad sequences to the longest sequence\n",
|
||||
" input_ids += [pad_token_id] * (max_length - len(input_ids))\n",
|
||||
" input_tensor = torch.tensor(input_ids, device=device).unsqueeze(0) # add batch dimension\n",
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user