File size: 1,303 Bytes
516a027 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 |
python main. py
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
model. safetensors.index.json: 100%|
| 13.5k/13.5k [00:00‹?, PB/s]
model-00001-of-00002. safetensors: 100%
| 4.95G/4.95G [07:27<00:00, 11. 1MB/s]
model-00002-of-00002. safetensors: 100%
67. 1M/67.1M [00:05<00:00, 11.5MB/s]
Downloading shards: 100% ||
| 2/2 [07:35‹00:00, 227.61s/it]
Gemma's activation function should be approximate GeLU and not exact GeLU. Changing the activation function to 'gelu_pytorch_tanh.if you want to use the legacy "gelu', edit the "model.config to
set hidden_activation=gelu*
instead of todden act
instead of hidden_act. See https://github.com/huggingface/transformers/pull/29402 for
more details.
Loading checkpoint shards: 100%|
| 2/2 [00:03<00:00, 1.87s/itl
generation_config json: 100%||
137/137[00:00<?」3B/s]
nexa model result:
a pouto using the specified caea and resolutiou stones iption: rame rs a photo (cama a):)
Captures
- camera (str): Specifies the camera
to use. Can be \'front\' or \'back\'. The default is \'back\'. \n\n
Returns: \n
- str: The string contains the file
2624 t 12 4a.
Photo if nees at ay 96 83662387968t, ample: /storage/emulated/o/Pictures/NAPP/3N
123456.Jpg\'\n latency: 367.85967230796814 |