File size: 510 Bytes
c446951 |
1 2 3 4 5 6 7 8 9 10 |
You can set up a server to use computer vision models with Inference on the following devices:
- ARM CPU (macOS, Raspberry Pi)
- x86 CPU (macOS, Linux, Windows)
- NVIDIA GPU
- NVIDIA Jetson (JetPack 4.5.x, JetPack 4.6.x, JetPack 5.x)
You can also run Inference in the cloud with the [Roboflow hosted inference offering](https://docs.roboflow.com/deploy/hosted-api).
To learn more about device compatability with different models, refer to the [model compatability matrix](/quickstart/compatability_matrix). |