Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
aws-neuron
/
optimum-neuron-cache
like
28
Follow
AWS Inferentia and Trainium
155
License:
apache-2.0
Model card
Files
Files and versions
xet
Community
636
main
optimum-neuron-cache
/
inference-cache-config
/
trn2
4.43 kB
4 contributors
History:
2 commits
dacorvo
HF Staff
use longer sequence length for llama3 on trn2
f8538f0
verified
about 6 hours ago
llama3.json
2.58 kB
use longer sequence length for llama3 on trn2
about 6 hours ago
llama4.json
1.1 kB
add trn2 cached configs subdirectory
3 months ago
qwen3-moe.json
751 Bytes
add trn2 cached configs subdirectory
3 months ago