Hugging Face
Models
Datasets
Spaces
Posts
Docs
Solutions
Pricing
Log In
Sign Up
aws-neuron
/
optimum-neuron-cache
like
10
Follow
AWS Inferentia and Trainium
53
License:
apache-2.0
Model card
Files
Files and versions
Community
235
a74d412
optimum-neuron-cache
/
inference-cache-config
7 contributors
History:
29 commits
Jingya
HF staff
Temporarily remove SD 1.5 from Runway
a74d412
verified
2 months ago
gpt2.json
398 Bytes
Add more gpt2 configurations
7 months ago
llama-variants.json
2.63 kB
Update inference-cache-config/llama-variants.json
4 months ago
llama2-70b.json
287 Bytes
Create llama2-70b.json
4 months ago
llama2-7b-13b.json
2.02 kB
Rename inference-cache-config/llama2.json to inference-cache-config/llama2-7b-13b.json
4 months ago
llama3-70b.json
283 Bytes
Create llama3-70b.json
4 months ago
llama3-8b.json
883 Bytes
Rename inference-cache-config/llama3.json to inference-cache-config/llama3-8b.json
4 months ago
mistral-variants.json
3.29 kB
Remove SalesForce embedding model
7 months ago
mistral.json
1.46 kB
Add more batch_size for mistral on smaller instances
5 months ago
mixtral.json
294 Bytes
Create mixtral.json
5 months ago
stable-diffusion.json
1.54 kB
Temporarily remove SD 1.5 from Runway
2 months ago