neuronbox, one yaml file between you and a running model.
Compatible across the stack you already use
One manifest. Hashed envs. Live dashboard. Everything else gets out of your way.
neuron.yaml ties your model, Python stack, and GPU expectations into a single source of truth. Same command on your laptop and your server.
Same manifest shape always resolves to the same env directory under the store. Change a package and only that environment rebuilds.
neuron pull fetches any HF-style org/model into ~/.neuronbox/store. Every project shares the same weights, with no 50 GB duplicates.
Live sessions, tok/s, VRAM estimates, and host GPU probes, all in one terminal view. neurond keeps state; neuron stats prints it. Session names, PIDs, and model paths stay aligned so you always know which run owns the GPU.
Throughput history is drawn client-side for a snappy ~10 Hz refresh without hammering the daemon. Quit with q or Esc when you are done exploring.
neuron swap updates daemon state and swap_signal.json. Your long-lived worker reacts immediately.
Add runtime.mode: oci to the manifest and neuron run --oci handles Docker mounts and the NVIDIA toolkit. Your call, not the default.
If you have an idea or a request for neuronbox, please get in touch.
Contact usOne install, one manifest, one command to hit your entrypoint.
Requires Rust. Also needs neurond beside the binary, or set NEUROND_PATH.
Describe your project once in neuron.yaml. NeuronBox handles the env, the weights, and the wiring.
Same manifest shape always resolves to the same venv. No rebuilds.
Set min_vram and NeuronBox warns you before the run, not during.
neuron run resolves, installs, wires env vars, and spawns your script.
model:source: hub # pulled from HF on first runname: mistralai/Mistral-7B-v0.1runtime:python: "3.11"packages:- transformers- accelerate- torchgpu:min_vram: 12 # soft check — warns, never blocksentrypoint: scripts/run_inference.py
neuronbox is not a Docker replacement. It's what you reach for before Docker becomes necessary.

neuronbox
for daily local ML work
Good for
Tradeoffs
Docker
for production parity and orchestration
Good for
Tradeoffs
When you need Docker, neuronbox has you covered, add runtime.mode: oci to your manifest and neuron run --oci handles the rest.
Sessions and GPU context in one terminal view. No polling scripts, no guessing.
try it yourself, neuron dashboard or neuron dashboard --demo