Instructions to use allenai/MolmoWeb-4B-Native with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use allenai/MolmoWeb-4B-Native with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("image-text-to-text", model="allenai/MolmoWeb-4B-Native")# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("allenai/MolmoWeb-4B-Native", dtype="auto") - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use allenai/MolmoWeb-4B-Native with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "allenai/MolmoWeb-4B-Native" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "allenai/MolmoWeb-4B-Native", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker
docker model run hf.co/allenai/MolmoWeb-4B-Native
- SGLang
How to use allenai/MolmoWeb-4B-Native with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "allenai/MolmoWeb-4B-Native" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "allenai/MolmoWeb-4B-Native", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "allenai/MolmoWeb-4B-Native" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "allenai/MolmoWeb-4B-Native", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }' - Docker Model Runner
How to use allenai/MolmoWeb-4B-Native with Docker Model Runner:
docker model run hf.co/allenai/MolmoWeb-4B-Native
Update README.md
#1
by peter-sushko - opened
README.md
CHANGED
|
@@ -53,4 +53,21 @@ Please refer to our [Github repo](https://github.com/allenai/molmoweb/) for infe
|
|
| 53 |
|
| 54 |
## License and Use
|
| 55 |
|
| 56 |
-
This model is licensed under Apache 2.0. It is intended for research and educational use in accordance with Ai2’s [Responsible Use Guidelines](https://allenai.org/responsible-use).
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 53 |
|
| 54 |
## License and Use
|
| 55 |
|
| 56 |
+
This model is licensed under Apache 2.0. It is intended for research and educational use in accordance with Ai2’s [Responsible Use Guidelines](https://allenai.org/responsible-use).
|
| 57 |
+
|
| 58 |
+
## Citation
|
| 59 |
+
|
| 60 |
+
If you use this dataset, please cite:
|
| 61 |
+
|
| 62 |
+
[arXiv:2604.08516](https://arxiv.org/abs/2604.08516)
|
| 63 |
+
|
| 64 |
+
```bibtex
|
| 65 |
+
@misc{gupta2026molmowebopenvisualweb,
|
| 66 |
+
title={MolmoWeb: Open Visual Web Agent and Open Data for the Open Web},
|
| 67 |
+
author={Tanmay Gupta and Piper Wolters and Zixian Ma and Peter Sushko and Rock Yuren Pang and Diego Llanes and Yue Yang and Taira Anderson and Boyuan Zheng and Zhongzheng Ren and Harsh Trivedi and Taylor Blanton and Caleb Ouellette and Winson Han and Ali Farhadi and Ranjay Krishna},
|
| 68 |
+
year={2026},
|
| 69 |
+
eprint={2604.08516},
|
| 70 |
+
archivePrefix={arXiv},
|
| 71 |
+
primaryClass={cs.CV},
|
| 72 |
+
url={https://arxiv.org/abs/2604.08516},
|
| 73 |
+
}
|