Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
aws-neuron
/
optimum-neuron-cache
like
32
Follow
AWS Inferentia and Trainium
172
License:
apache-2.0
Model card
Files
Files and versions
xet
Community
676
5116316
optimum-neuron-cache
/
inference-cache-config
16 kB
Ctrl+K
Ctrl+K
5 contributors
History:
64 commits
Jingya
HF Staff
add pixart and remove deprecated
e5f06c7
verified
10 months ago
diffusion.json
2.31 kB
add pixart and remove deprecated
10 months ago
gpt2.json
Safe
398 Bytes
Add more gpt2 configurations
about 2 years ago
granite.json
Safe
1.3 kB
Add configuration for granite models
over 1 year ago
llama-variants.json
Safe
1.45 kB
Add DeepSeek distilled versions of LLama 8B
about 1 year ago
llama.json
1.84 kB
Added TinyLlama as requested by Jim burtoft
12 months ago
llama2-70b.json
Safe
287 Bytes
Create llama2-70b.json
almost 2 years ago
llama3-70b.json
Safe
584 Bytes
Add DeepSeek distilled model
about 1 year ago
llama3.1-70b.json
Safe
289 Bytes
Rename inference-cache-config/Llama3.1-70b.json to inference-cache-config/llama3.1-70b.json
over 1 year ago
mistral-variants.json
Safe
1.04 kB
Remove obsolete mistral variants
over 1 year ago
mistral.json
Safe
1.87 kB
Update inference-cache-config/mistral.json
over 1 year ago
mixtral.json
Safe
583 Bytes
Update inference-cache-config/mixtral.json
over 1 year ago
phi4.json
Safe
556 Bytes
Add phi4 cached configurations
about 1 year ago
qwen2.5-large.json
Safe
849 Bytes
Update inference-cache-config/qwen2.5-large.json
about 1 year ago
qwen2.5.json
Safe
2.69 kB
Add DeepSeek distilled models
about 1 year ago