Your hardware
Run models on your own devices for fast experimentation.
The guides in this section walk you through compiling, profiling, and running inference on your own devices — a Raspberry Pi, NVIDIA Jetson board, desktop GPU, or any Linux machine reachable over SSH. This is ideal when you want:
- Fast feedback loops — jobs run directly on the target device. With ONNX Runtime, a full compile-and-profile cycle for a MobileNetV2 finishes in under 20 seconds — compared to ~10 minutes for the same model on a cloud provider.
- Full control — use any device you have, not just those available in a managed cloud.
- No cloud dependency — all computation stays on your network.
If you need to test across many different edge devices without purchasing and managing them, see the Cloud guides instead — cloud providers give you access to a wide range of hardware from your laptop.
Interested in a private device farm on your own infrastructure?
Keep your data on-site and experiment at scale.
Prerequisites
Before following any guide in this section, make sure you have completed the setup guide to:
- Create an Embedl Hub account
- Install the
embedl-hubPython library - Configure an API key
SSH access
Your device must be reachable over SSH with passwordless authentication. If that’s not set up yet:
Generate an SSH key pair (if you haven’t already):
ssh-keygen -t ed25519Copy the public key to the target device:
ssh-copy-id user@192.168.1.42Verify that you can connect without a password:
ssh user@192.168.1.42 echo "Connected!"
Available guides
- ONNX Runtime — compile, profile,
and invoke ONNX Runtime models on any Linux device using
embedl-onnxruntime. - TensorRT — compile, profile,
and invoke TensorRT models on NVIDIA GPU-equipped devices using
trtexec.