LLMs-from-scratch/pyproject.toml
Sebastian Raschka 2dc2df593a Llama3Fast (#593)
* Llama3Fast

* Update pkg/llms_from_scratch/tests/test_llama3.py
2025-04-01 12:56:11 -05:00

53 lines
1002 B
TOML

[build-system]
requires = ["setuptools>=61.0", "wheel"]
build-backend = "setuptools.build_meta"
[project]
name = "llms-from-scratch"
version = "1.0.6"
description = "Implement a ChatGPT-like LLM in PyTorch from scratch, step by step"
readme = "README.md"
requires-python = ">=3.10"
dependencies = [
"torch>=2.3.0",
"jupyterlab>=4.0",
"tiktoken>=0.5.1",
"matplotlib>=3.7.1",
"tensorflow>=2.18.0",
"tqdm>=4.66.1",
"numpy>=1.26,<2.1",
"pandas>=2.2.1",
"pip>=25.0.1",
"pytest>=8.3.5",
]
[tool.uv.sources]
llms-from-scratch = { workspace = true }
[dependency-groups]
dev = [
"build>=1.2.2.post1",
"llms-from-scratch",
"twine>=6.1.0",
]
[tool.ruff]
line-length = 140
[tool.ruff.lint]
exclude = [".venv"]
# Ignored rules (W504 removed)
ignore = [
"C406", "E226", "E402", "E702", "E703",
"E722", "E731", "E741"
]
# `llms_from_scratch` PyPI package
[tool.setuptools]
package-dir = {"" = "pkg"}
[tool.setuptools.packages.find]
where = ["pkg"]