nanochat/services/inference/pyproject.toml
2026-04-16 11:19:18 -07:00

39 lines
811 B
TOML

[project]
name = "nanochat-inference-service"
version = "0.1.0"
description = "Standalone FastAPI inference service for nanochat"
readme = "README.md"
requires-python = ">=3.12"
dependencies = [
"fastapi>=0.117.1",
"filelock>=3.18.0",
"huggingface_hub>=0.35.0",
"pydantic-settings>=2.10.1",
"requests>=2.32.0",
"rustbpe>=0.1.0",
"tiktoken>=0.11.0",
"tokenizers>=0.22.0",
"torch==2.9.1",
"transformers>=4.57.3",
"uvicorn>=0.36.0",
]
[dependency-groups]
dev = [
"httpx>=0.28.1",
"pytest>=8.0.0",
]
[tool.pytest.ini_options]
testpaths = ["tests"]
python_files = ["test_*.py"]
python_functions = ["test_*"]
[tool.uv.sources]
torch = { index = "pytorch-cpu" }
[[tool.uv.index]]
name = "pytorch-cpu"
url = "https://download.pytorch.org/whl/cpu"
explicit = true