Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
aws-neuron
/
optimum-neuron-cache
like
13
Follow
AWS Inferentia and Trainium
76
License:
apache-2.0
Model card
Files
Files and versions
Community
300
d8e677c
optimum-neuron-cache
/
inference-cache-config
8 contributors
History:
54 commits
dacorvo
HF staff
Add configuration for granite models
687da09
verified
29 days ago
gpt2.json
Safe
398 Bytes
Add more gpt2 configurations
10 months ago
granite.json
Safe
1.3 kB
Add configuration for granite models
29 days ago
llama-variants.json
Safe
559 Bytes
Remove obsolete llama variants
4 months ago
llama.json
Safe
1.67 kB
Update inference-cache-config/llama.json
4 months ago
llama2-70b.json
Safe
287 Bytes
Create llama2-70b.json
7 months ago
llama3-70b.json
Safe
283 Bytes
Update inference-cache-config/llama3-70b.json
4 months ago
llama3.1-70b.json
Safe
289 Bytes
Rename inference-cache-config/Llama3.1-70b.json to inference-cache-config/llama3.1-70b.json
4 months ago
mistral-variants.json
Safe
1.04 kB
Remove obsolete mistral variants
4 months ago
mistral.json
Safe
1.8 kB
Update inference-cache-config/mistral.json
4 months ago
mixtral.json
Safe
583 Bytes
Update inference-cache-config/mixtral.json
4 months ago
qwen2.5-large.json
Safe
558 Bytes
Rename inference-cache-config/qwen-2.5-large.json to inference-cache-config/qwen2.5-large.json
about 2 months ago
qwen2.5.json
Safe
1.45 kB
Rename inference-cache-config/qwen2.5 to inference-cache-config/qwen2.5.json
about 2 months ago
stable-diffusion.json
Safe
1.91 kB
Update inference-cache-config/stable-diffusion.json
4 months ago