Deployment
Hugging Face Spaces demo
This repository includes a Docker-based Hugging Face Space configuration for the interactive Streamlit app:
The Space metadata lives at the top of README.md and points Hugging Face to
the root Dockerfile. The container installs the package, starts Ollama, pulls
the models required by the bundled Ollama configs, and then serves Streamlit on
port 7860.
Prepared Ollama models:
deepseek-r1:latestqwen3:8b
The startup script is deploy/huggingface/start.sh. In the Hugging Face Space
settings, you can override the models pulled at boot with the
SPACE_OLLAMA_MODELS environment variable, for example:
SPACE_OLLAMA_MODELS="deepseek-r1:latest"
On Hugging Face’s free CPU tier, expect cold starts and slow inference. The configuration is intended as a convenient public demo, not as a production serving setup.
GitHub Pages documentation
The repository includes a GitHub Actions workflow at .github/workflows/docs.yml that builds the Sphinx documentation and deploys it to GitHub Pages.
The docs build is driven from source files in docs/. The generated docs/_build directory should remain out of version control.