Install:
cargo install --git https://github.com/hellas-ai/nodeExecute:
cargo run -- execute -p heyExecute locally with the catgrad backend:
cargo run -- execute --local -p heyLocal execution uses the same catgrad executor backend as serve and prefers
accelerated backends when available (Metal on macOS, --features cuda on Linux).
Verify a remote execution against the local catgrad backend:
cargo run -- execute --verify-local -p heyInstall server features:
cargo install --git https://github.com/hellas-ai/node --features serveRun server:
hellas-cli serve --download-policy=eager --execute-policy=eager
Node Address: bb18ebc065d836ecc7e1f33972d2c17eac9894cd33ce4916f66cb1165ccc7550
RPC server running. Press Ctrl+C to stophellas-cli serve without policy flags now starts in deny-by-default mode
(--download-policy=skip --execute-policy=skip). Only pass eager or allow-list
policies when you intentionally want a node to serve remote work.
Preload weights on startup:
hellas-cli serve \
--download-policy=eager \
--execute-policy=eager \
--preload HuggingFaceTB/SmolLM2-135M-InstructRepeat --preload to warm multiple models before the node starts serving.
Run client:
cargo run -- execute bb18ebc065d836ecc7e1f33972d2c17eac9894cd33ce4916f66cb1165ccc7550 -p hey
Hello! How can I help you today?Monitor discovery and peer health:
cargo run -- monitor --timeout-secs 30Run HTTP gateway (OpenAI / Anthropic / plain completions over Hellas network):
cargo run -- gateway --port 8080Routes:
POST /v1/chat/completions
POST /v1/messages
POST /v1/completionsDocker images: .#docker-cpu, .#docker-cuda12-sm89, etc. They stream to stdout.
$(nix build .#docker-cuda12-sm89 --print-out-paths) | docker load
nix run .#docker-push-all # push all images to ghcr.io/hellas-ai/nodeRun a CUDA server with persistent HF cache, metrics, and Jaeger tracing:
docker run --rm -it \
--device=nvidia.com/gpu=all \
-p 31145:31145/udp \
-p 9090:9090 \
-v ~/.cache/huggingface:/home/hellas/.cache/huggingface \
-e OTEL_EXPORTER_OTLP_TRACES_ENDPOINT=http://jaeger:4318/v1/traces \
ghcr.io/hellas-ai/node:cuda12-sm89 \
--download-policy=eager --execute-policy=eager \
--metrics-port=9090 \
--preload HuggingFaceTB/SmolLM2-135M-InstructAvailable in the dev shell (nix develop):
cargo audit # security advisories
cargo outdated --workspace --root-deps-only # outdated deps
cargo update --workspace # update Cargo.lock