Missing Docker image and MI300X
#2
by
twinsen123
- opened
Two questions for using this model
- The
rocm/vllm-private:vllm_dev_base_mxfp4_20260122image specified model card is not accessible, could you update the model card which VLLM is compatible to this quantization - It is possible to run this model using MI300X?
- In this docker images we use this vllm commit
74c583bc508c2dafb9e95bab3b635884e4a021f3and modify the following code.
https://github.com/vllm-project/vllm/blob/main/vllm/model_executor/models/glm4_moe.py
- We haven't tested it on the mi300, you can give it a try.