Update inference-cache-config/llama.json
Browse files
inference-cache-config/llama.json
CHANGED
@@ -42,5 +42,21 @@
|
|
42 |
"num_cores": 8,
|
43 |
"auto_cast_type": "bf16"
|
44 |
}
|
45 |
-
]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
46 |
}
|
|
|
42 |
"num_cores": 8,
|
43 |
"auto_cast_type": "bf16"
|
44 |
}
|
45 |
+
],
|
46 |
+
"meta-llama/Llama-2-7b-hf": [
|
47 |
+
{
|
48 |
+
"batch_size": 1,
|
49 |
+
"sequence_length": 2048,
|
50 |
+
"num_cores": 2,
|
51 |
+
"auto_cast_type": "bf16"
|
52 |
+
}
|
53 |
+
],
|
54 |
+
"meta-llama/Llama-2-13b-hf": [
|
55 |
+
{
|
56 |
+
"batch_size": 1,
|
57 |
+
"sequence_length": 2048,
|
58 |
+
"num_cores": 8,
|
59 |
+
"auto_cast_type": "bf16"
|
60 |
+
}
|
61 |
+
],
|
62 |
}
|