Text Generation
Transformers
PyTorch
t5
text2text-generation
biology
single-cell
single-cell analysis
text-generation-inference
Instructions to use zjunlp/chatcell-large with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use zjunlp/chatcell-large with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="zjunlp/chatcell-large")# Load model directly from transformers import AutoTokenizer, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("zjunlp/chatcell-large") model = AutoModelForSeq2SeqLM.from_pretrained("zjunlp/chatcell-large") - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use zjunlp/chatcell-large with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "zjunlp/chatcell-large" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "zjunlp/chatcell-large", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker
docker model run hf.co/zjunlp/chatcell-large
- SGLang
How to use zjunlp/chatcell-large with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "zjunlp/chatcell-large" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "zjunlp/chatcell-large", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "zjunlp/chatcell-large" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "zjunlp/chatcell-large", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }' - Docker Model Runner
How to use zjunlp/chatcell-large with Docker Model Runner:
docker model run hf.co/zjunlp/chatcell-large
Commit History
Update README.md 3faf080 verified
Update README.md e1bf81a verified
Update README.md bd19b9e verified
Update README.md 642027d verified
Update README.md f15df98 verified
Update README.md 7ce53b6 verified
Create handler.py 5e439c3 verified
Update README.md 6354e14 verified
Yin Fang commited on
Update README.md c54400c verified
Update README.md 3d14fc8 verified
Yin Fang commited on
Upload 8 files 899cacc verified
Yin Fang commited on
Upload 8 files 5b55003 verified
Yin Fang commited on
initial commit cca7fbc verified
Yin Fang commited on