mirror of
https://github.com/rasbt/LLMs-from-scratch.git
synced 2025-12-01 01:20:33 +00:00
prevent self.apply_chat_template being applied for base Qwen models
This commit is contained in:
parent
8add26cbe9
commit
df504397a8
@ -973,7 +973,7 @@
|
|||||||
},
|
},
|
||||||
{
|
{
|
||||||
"cell_type": "code",
|
"cell_type": "code",
|
||||||
"execution_count": 18,
|
"execution_count": null,
|
||||||
"id": "b68ab489-48e5-471e-a814-56cda2d60f81",
|
"id": "b68ab489-48e5-471e-a814-56cda2d60f81",
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"outputs": [],
|
"outputs": [],
|
||||||
@ -1000,6 +1000,7 @@
|
|||||||
" self.apply_chat_template = apply_chat_template\n",
|
" self.apply_chat_template = apply_chat_template\n",
|
||||||
" self.add_generation_prompt = add_generation_prompt\n",
|
" self.add_generation_prompt = add_generation_prompt\n",
|
||||||
" self.add_thinking = add_thinking\n",
|
" self.add_thinking = add_thinking\n",
|
||||||
|
" self.repo_id=repo_id\n",
|
||||||
"\n",
|
"\n",
|
||||||
" tok_file = Path(tokenizer_file_path)\n",
|
" tok_file = Path(tokenizer_file_path)\n",
|
||||||
" self._tok = Tokenizer.from_file(str(tok_file))\n",
|
" self._tok = Tokenizer.from_file(str(tok_file))\n",
|
||||||
@ -1012,7 +1013,7 @@
|
|||||||
" self.pad_token_id = self._special_to_id[\"<|endoftext|>\"]\n",
|
" self.pad_token_id = self._special_to_id[\"<|endoftext|>\"]\n",
|
||||||
" self.eos_token_id = self.pad_token_id\n",
|
" self.eos_token_id = self.pad_token_id\n",
|
||||||
"\n",
|
"\n",
|
||||||
" if repo_id and \"Base\" not in repo_id:\n",
|
" if self.repo_id and \"Base\" not in self.repo_id:\n",
|
||||||
" eos_token = \"<|im_end|>\"\n",
|
" eos_token = \"<|im_end|>\"\n",
|
||||||
" else:\n",
|
" else:\n",
|
||||||
" eos_token = \"<|endoftext|>\"\n",
|
" eos_token = \"<|endoftext|>\"\n",
|
||||||
@ -1020,7 +1021,7 @@
|
|||||||
" self.eos_token_id = self._special_to_id[eos_token]\n",
|
" self.eos_token_id = self._special_to_id[eos_token]\n",
|
||||||
"\n",
|
"\n",
|
||||||
" def encode(self, text, chat_wrapped=None):\n",
|
" def encode(self, text, chat_wrapped=None):\n",
|
||||||
" if chat_wrapped is None:\n",
|
" if chat_wrapped is None and self.repo_id and \"Base\" not in self.repo_id:\n",
|
||||||
" chat_wrapped = self.apply_chat_template\n",
|
" chat_wrapped = self.apply_chat_template\n",
|
||||||
"\n",
|
"\n",
|
||||||
" stripped = text.strip()\n",
|
" stripped = text.strip()\n",
|
||||||
@ -1239,7 +1240,7 @@
|
|||||||
"provenance": []
|
"provenance": []
|
||||||
},
|
},
|
||||||
"kernelspec": {
|
"kernelspec": {
|
||||||
"display_name": "Python 3 (ipykernel)",
|
"display_name": ".venv",
|
||||||
"language": "python",
|
"language": "python",
|
||||||
"name": "python3"
|
"name": "python3"
|
||||||
},
|
},
|
||||||
@ -1253,7 +1254,7 @@
|
|||||||
"name": "python",
|
"name": "python",
|
||||||
"nbconvert_exporter": "python",
|
"nbconvert_exporter": "python",
|
||||||
"pygments_lexer": "ipython3",
|
"pygments_lexer": "ipython3",
|
||||||
"version": "3.10.16"
|
"version": "3.12.6"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"nbformat": 4,
|
"nbformat": 4,
|
||||||
|
|||||||
@ -915,7 +915,7 @@
|
|||||||
},
|
},
|
||||||
{
|
{
|
||||||
"cell_type": "code",
|
"cell_type": "code",
|
||||||
"execution_count": 17,
|
"execution_count": null,
|
||||||
"id": "b68ab489-48e5-471e-a814-56cda2d60f81",
|
"id": "b68ab489-48e5-471e-a814-56cda2d60f81",
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"outputs": [],
|
"outputs": [],
|
||||||
@ -942,6 +942,7 @@
|
|||||||
" self.apply_chat_template = apply_chat_template\n",
|
" self.apply_chat_template = apply_chat_template\n",
|
||||||
" self.add_generation_prompt = add_generation_prompt\n",
|
" self.add_generation_prompt = add_generation_prompt\n",
|
||||||
" self.add_thinking = add_thinking\n",
|
" self.add_thinking = add_thinking\n",
|
||||||
|
" self.repo_id=repo_id\n",
|
||||||
"\n",
|
"\n",
|
||||||
" tok_file = Path(tokenizer_file_path)\n",
|
" tok_file = Path(tokenizer_file_path)\n",
|
||||||
" self._tok = Tokenizer.from_file(str(tok_file))\n",
|
" self._tok = Tokenizer.from_file(str(tok_file))\n",
|
||||||
@ -954,7 +955,7 @@
|
|||||||
" self.pad_token_id = self._special_to_id[\"<|endoftext|>\"]\n",
|
" self.pad_token_id = self._special_to_id[\"<|endoftext|>\"]\n",
|
||||||
" self.eos_token_id = self.pad_token_id\n",
|
" self.eos_token_id = self.pad_token_id\n",
|
||||||
"\n",
|
"\n",
|
||||||
" if repo_id and \"Base\" not in repo_id:\n",
|
" if self.repo_id and \"Base\" not in self.repo_id:\n",
|
||||||
" eos_token = \"<|im_end|>\"\n",
|
" eos_token = \"<|im_end|>\"\n",
|
||||||
" else:\n",
|
" else:\n",
|
||||||
" eos_token = \"<|endoftext|>\"\n",
|
" eos_token = \"<|endoftext|>\"\n",
|
||||||
@ -962,7 +963,7 @@
|
|||||||
" self.eos_token_id = self._special_to_id[eos_token]\n",
|
" self.eos_token_id = self._special_to_id[eos_token]\n",
|
||||||
"\n",
|
"\n",
|
||||||
" def encode(self, text, chat_wrapped=None):\n",
|
" def encode(self, text, chat_wrapped=None):\n",
|
||||||
" if chat_wrapped is None:\n",
|
" if chat_wrapped is None and self.repo_id and \"Base\" not in self.repo_id:\n",
|
||||||
" chat_wrapped = self.apply_chat_template\n",
|
" chat_wrapped = self.apply_chat_template\n",
|
||||||
"\n",
|
"\n",
|
||||||
" stripped = text.strip()\n",
|
" stripped = text.strip()\n",
|
||||||
@ -1221,7 +1222,7 @@
|
|||||||
"provenance": []
|
"provenance": []
|
||||||
},
|
},
|
||||||
"kernelspec": {
|
"kernelspec": {
|
||||||
"display_name": "Python 3 (ipykernel)",
|
"display_name": ".venv",
|
||||||
"language": "python",
|
"language": "python",
|
||||||
"name": "python3"
|
"name": "python3"
|
||||||
},
|
},
|
||||||
@ -1235,7 +1236,7 @@
|
|||||||
"name": "python",
|
"name": "python",
|
||||||
"nbconvert_exporter": "python",
|
"nbconvert_exporter": "python",
|
||||||
"pygments_lexer": "ipython3",
|
"pygments_lexer": "ipython3",
|
||||||
"version": "3.10.16"
|
"version": "3.12.6"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"nbformat": 4,
|
"nbformat": 4,
|
||||||
|
|||||||
@ -960,7 +960,7 @@
|
|||||||
},
|
},
|
||||||
{
|
{
|
||||||
"cell_type": "code",
|
"cell_type": "code",
|
||||||
"execution_count": 20,
|
"execution_count": null,
|
||||||
"id": "b68ab489-48e5-471e-a814-56cda2d60f81",
|
"id": "b68ab489-48e5-471e-a814-56cda2d60f81",
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"outputs": [],
|
"outputs": [],
|
||||||
@ -987,6 +987,7 @@
|
|||||||
" self.apply_chat_template = apply_chat_template\n",
|
" self.apply_chat_template = apply_chat_template\n",
|
||||||
" self.add_generation_prompt = add_generation_prompt\n",
|
" self.add_generation_prompt = add_generation_prompt\n",
|
||||||
" self.add_thinking = add_thinking\n",
|
" self.add_thinking = add_thinking\n",
|
||||||
|
" self.repo_id=repo_id\n",
|
||||||
"\n",
|
"\n",
|
||||||
" tok_file = Path(tokenizer_file_path)\n",
|
" tok_file = Path(tokenizer_file_path)\n",
|
||||||
" self._tok = Tokenizer.from_file(str(tok_file))\n",
|
" self._tok = Tokenizer.from_file(str(tok_file))\n",
|
||||||
@ -999,7 +1000,7 @@
|
|||||||
" self.pad_token_id = self._special_to_id[\"<|endoftext|>\"]\n",
|
" self.pad_token_id = self._special_to_id[\"<|endoftext|>\"]\n",
|
||||||
" self.eos_token_id = self.pad_token_id\n",
|
" self.eos_token_id = self.pad_token_id\n",
|
||||||
"\n",
|
"\n",
|
||||||
" if repo_id and \"Base\" not in repo_id:\n",
|
" if self.repo_id and \"Base\" not in self.repo_id:\n",
|
||||||
" eos_token = \"<|im_end|>\"\n",
|
" eos_token = \"<|im_end|>\"\n",
|
||||||
" else:\n",
|
" else:\n",
|
||||||
" eos_token = \"<|endoftext|>\"\n",
|
" eos_token = \"<|endoftext|>\"\n",
|
||||||
@ -1007,7 +1008,7 @@
|
|||||||
" self.eos_token_id = self._special_to_id[eos_token]\n",
|
" self.eos_token_id = self._special_to_id[eos_token]\n",
|
||||||
"\n",
|
"\n",
|
||||||
" def encode(self, text, chat_wrapped=None):\n",
|
" def encode(self, text, chat_wrapped=None):\n",
|
||||||
" if chat_wrapped is None:\n",
|
" if chat_wrapped is None and self.repo_id and \"Base\" not in self.repo_id:\n",
|
||||||
" chat_wrapped = self.apply_chat_template\n",
|
" chat_wrapped = self.apply_chat_template\n",
|
||||||
"\n",
|
"\n",
|
||||||
" stripped = text.strip()\n",
|
" stripped = text.strip()\n",
|
||||||
@ -1206,7 +1207,7 @@
|
|||||||
"provenance": []
|
"provenance": []
|
||||||
},
|
},
|
||||||
"kernelspec": {
|
"kernelspec": {
|
||||||
"display_name": "Python 3 (ipykernel)",
|
"display_name": ".venv",
|
||||||
"language": "python",
|
"language": "python",
|
||||||
"name": "python3"
|
"name": "python3"
|
||||||
},
|
},
|
||||||
@ -1220,7 +1221,7 @@
|
|||||||
"name": "python",
|
"name": "python",
|
||||||
"nbconvert_exporter": "python",
|
"nbconvert_exporter": "python",
|
||||||
"pygments_lexer": "ipython3",
|
"pygments_lexer": "ipython3",
|
||||||
"version": "3.10.16"
|
"version": "3.12.6"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"nbformat": 4,
|
"nbformat": 4,
|
||||||
|
|||||||
@ -902,7 +902,7 @@
|
|||||||
},
|
},
|
||||||
{
|
{
|
||||||
"cell_type": "code",
|
"cell_type": "code",
|
||||||
"execution_count": 19,
|
"execution_count": null,
|
||||||
"id": "b68ab489-48e5-471e-a814-56cda2d60f81",
|
"id": "b68ab489-48e5-471e-a814-56cda2d60f81",
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"outputs": [],
|
"outputs": [],
|
||||||
@ -929,6 +929,7 @@
|
|||||||
" self.apply_chat_template = apply_chat_template\n",
|
" self.apply_chat_template = apply_chat_template\n",
|
||||||
" self.add_generation_prompt = add_generation_prompt\n",
|
" self.add_generation_prompt = add_generation_prompt\n",
|
||||||
" self.add_thinking = add_thinking\n",
|
" self.add_thinking = add_thinking\n",
|
||||||
|
" self.repo_id=repo_id\n",
|
||||||
"\n",
|
"\n",
|
||||||
" tok_file = Path(tokenizer_file_path)\n",
|
" tok_file = Path(tokenizer_file_path)\n",
|
||||||
" self._tok = Tokenizer.from_file(str(tok_file))\n",
|
" self._tok = Tokenizer.from_file(str(tok_file))\n",
|
||||||
@ -941,7 +942,7 @@
|
|||||||
" self.pad_token_id = self._special_to_id[\"<|endoftext|>\"]\n",
|
" self.pad_token_id = self._special_to_id[\"<|endoftext|>\"]\n",
|
||||||
" self.eos_token_id = self.pad_token_id\n",
|
" self.eos_token_id = self.pad_token_id\n",
|
||||||
"\n",
|
"\n",
|
||||||
" if repo_id and \"Base\" not in repo_id:\n",
|
" if self.repo_id and \"Base\" not in self.repo_id:\n",
|
||||||
" eos_token = \"<|im_end|>\"\n",
|
" eos_token = \"<|im_end|>\"\n",
|
||||||
" else:\n",
|
" else:\n",
|
||||||
" eos_token = \"<|endoftext|>\"\n",
|
" eos_token = \"<|endoftext|>\"\n",
|
||||||
@ -949,7 +950,7 @@
|
|||||||
" self.eos_token_id = self._special_to_id[eos_token]\n",
|
" self.eos_token_id = self._special_to_id[eos_token]\n",
|
||||||
"\n",
|
"\n",
|
||||||
" def encode(self, text, chat_wrapped=None):\n",
|
" def encode(self, text, chat_wrapped=None):\n",
|
||||||
" if chat_wrapped is None:\n",
|
" if chat_wrapped is None and self.repo_id and \"Base\" not in self.repo_id:\n",
|
||||||
" chat_wrapped = self.apply_chat_template\n",
|
" chat_wrapped = self.apply_chat_template\n",
|
||||||
"\n",
|
"\n",
|
||||||
" stripped = text.strip()\n",
|
" stripped = text.strip()\n",
|
||||||
@ -1141,7 +1142,7 @@
|
|||||||
"provenance": []
|
"provenance": []
|
||||||
},
|
},
|
||||||
"kernelspec": {
|
"kernelspec": {
|
||||||
"display_name": "Python 3 (ipykernel)",
|
"display_name": ".venv",
|
||||||
"language": "python",
|
"language": "python",
|
||||||
"name": "python3"
|
"name": "python3"
|
||||||
},
|
},
|
||||||
@ -1155,7 +1156,7 @@
|
|||||||
"name": "python",
|
"name": "python",
|
||||||
"nbconvert_exporter": "python",
|
"nbconvert_exporter": "python",
|
||||||
"pygments_lexer": "ipython3",
|
"pygments_lexer": "ipython3",
|
||||||
"version": "3.10.16"
|
"version": "3.12.6"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"nbformat": 4,
|
"nbformat": 4,
|
||||||
|
|||||||
@ -531,6 +531,7 @@ class Qwen3Tokenizer:
|
|||||||
self.apply_chat_template = apply_chat_template
|
self.apply_chat_template = apply_chat_template
|
||||||
self.add_generation_prompt = add_generation_prompt
|
self.add_generation_prompt = add_generation_prompt
|
||||||
self.add_thinking = add_thinking
|
self.add_thinking = add_thinking
|
||||||
|
self.repo_id=repo_id
|
||||||
|
|
||||||
tok_file = Path(tokenizer_file_path)
|
tok_file = Path(tokenizer_file_path)
|
||||||
if not tok_file.is_file() and repo_id:
|
if not tok_file.is_file() and repo_id:
|
||||||
@ -549,7 +550,7 @@ class Qwen3Tokenizer:
|
|||||||
self.pad_token_id = self._special_to_id["<|endoftext|>"]
|
self.pad_token_id = self._special_to_id["<|endoftext|>"]
|
||||||
self.eos_token_id = self.pad_token_id
|
self.eos_token_id = self.pad_token_id
|
||||||
|
|
||||||
if repo_id and "Base" not in repo_id:
|
if self.repo_id and "Base" not in self.repo_id:
|
||||||
eos_token = "<|im_end|>"
|
eos_token = "<|im_end|>"
|
||||||
else:
|
else:
|
||||||
eos_token = "<|endoftext|>"
|
eos_token = "<|endoftext|>"
|
||||||
@ -557,7 +558,7 @@ class Qwen3Tokenizer:
|
|||||||
self.eos_token_id = self._special_to_id[eos_token]
|
self.eos_token_id = self._special_to_id[eos_token]
|
||||||
|
|
||||||
def encode(self, text, chat_wrapped=None):
|
def encode(self, text, chat_wrapped=None):
|
||||||
if chat_wrapped is None:
|
if chat_wrapped is None and self.repo_id and "Base" not in self.repo_id:
|
||||||
chat_wrapped = self.apply_chat_template
|
chat_wrapped = self.apply_chat_template
|
||||||
|
|
||||||
stripped = text.strip()
|
stripped = text.strip()
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user