You can set up a server to use computer vision models with Inference on the following devices: - ARM CPU (macOS, Raspberry Pi) - x86 CPU (macOS, Linux, Windows) - NVIDIA GPU - NVIDIA Jetson (JetPack 4.5.x, JetPack 4.6.x, JetPack 5.x) You can also run Inference in the cloud with the [Roboflow hosted inference offering](https://docs.roboflow.com/deploy/hosted-api). To learn more about device compatability with different models, refer to the [model compatability matrix](/quickstart/compatability_matrix).