Text Generation
Transformers
Safetensors
Dream
feature-extraction
diffusion
fast-inference
d3llm
conversational
custom_code
Instructions to use d3LLM/d3LLM_Dream with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use d3LLM/d3LLM_Dream with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="d3LLM/d3LLM_Dream", trust_remote_code=True) messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("d3LLM/d3LLM_Dream", trust_remote_code=True, dtype="auto") - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use d3LLM/d3LLM_Dream with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "d3LLM/d3LLM_Dream" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "d3LLM/d3LLM_Dream", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'Use Docker
docker model run hf.co/d3LLM/d3LLM_Dream
- SGLang
How to use d3LLM/d3LLM_Dream with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "d3LLM/d3LLM_Dream" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "d3LLM/d3LLM_Dream", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "d3LLM/d3LLM_Dream" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "d3LLM/d3LLM_Dream", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }' - Docker Model Runner
How to use d3LLM/d3LLM_Dream with Docker Model Runner:
docker model run hf.co/d3LLM/d3LLM_Dream
Commit History
Update README.md caa0744 verified
Update README.md dc81523 verified
Update README.md dd6c86a verified
Update README.md 1fd6f8c verified
Update README.md 11eeeeb verified
Update README.md 6b285cd verified
Update README.md c651d67 verified
Update README.md 41c9a13 verified
Update README.md 90cfadd verified
Create README.md d3d380d verified
Upload merges.txt with huggingface_hub b28b3f8 verified
Chien commited on
Upload vocab.json with huggingface_hub e48bc20 verified
Chien commited on
Upload special_tokens_map.json with huggingface_hub 822d99c verified
Chien commited on
Upload added_tokens.json with huggingface_hub bbcfbf4 verified
Chien commited on
Upload tokenizer_config.json with huggingface_hub ef1338e verified
Chien commited on
Upload model.safetensors.index.json with huggingface_hub 60ffcf1 verified
Chien commited on
Upload model-00004-of-00004.safetensors with huggingface_hub 00d7d60 verified
Chien commited on
Upload model-00003-of-00004.safetensors with huggingface_hub ddf2124 verified
Chien commited on
Upload model-00002-of-00004.safetensors with huggingface_hub cae62ac verified
Chien commited on
Upload model-00001-of-00004.safetensors with huggingface_hub 87f9400 verified
Chien commited on
Upload generation_config.json with huggingface_hub a4e1553 verified
Chien commited on
Upload config.json with huggingface_hub ccf134f verified
Chien commited on
initial commit c98fa69 verified
Chien commited on