Text Generation
Transformers
llamafile
code
granite
Eval Results
Inference Endpoints
jartine commited on
Commit
de1686b
1 Parent(s): 3048f7d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -265,7 +265,8 @@ Command template:
265
  The maximum context size of this model is 8192 tokens. These llamafiles
266
  use a default context size of 512 tokens. Whenever you need the maximum
267
  context size to be available with llamafile for any given model, you can
268
- pass the `-c 0` flag.
 
269
 
270
  ## About Quantization
271
 
 
265
  The maximum context size of this model is 8192 tokens. These llamafiles
266
  use a default context size of 512 tokens. Whenever you need the maximum
267
  context size to be available with llamafile for any given model, you can
268
+ pass the `-c 0` flag. The default temperature for these llamafiles is 0.
269
+ It can be changed, e.g. `--temp 0.8`.
270
 
271
  ## About Quantization
272