Synchronizing local compiler cache.
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +8 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_2354da4d82d493971ffa+2c2d707e/compile_flags.txt +1 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_2354da4d82d493971ffa+2c2d707e/model.hlo_module.pb +3 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_2354da4d82d493971ffa+2c2d707e/model.neff +3 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_2f1b54731e540b7a1e39+2c2d707e/compile_flags.txt +1 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_2f1b54731e540b7a1e39+2c2d707e/model.hlo_module.pb +3 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_2f1b54731e540b7a1e39+2c2d707e/model.neff +3 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_45c97e1b335e74b7fcf4+2c2d707e/compile_flags.txt +1 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_45c97e1b335e74b7fcf4+2c2d707e/model.hlo_module.pb +3 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_45c97e1b335e74b7fcf4+2c2d707e/model.neff +3 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_58ddfb26c2d4844ddc48+2c2d707e/compile_flags.txt +1 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_58ddfb26c2d4844ddc48+2c2d707e/model.hlo_module.pb +3 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_58ddfb26c2d4844ddc48+2c2d707e/model.neff +3 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_94243e60d465b93319ce+2c2d707e/compile_flags.txt +1 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_94243e60d465b93319ce+2c2d707e/model.hlo_module.pb +3 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_a6dc5c5ed3cdf7cd3eca+2c2d707e/compile_flags.txt +1 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_a6dc5c5ed3cdf7cd3eca+2c2d707e/model.hlo_module.pb +3 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_b4fff7ba577a27e2c997+2c2d707e/compile_flags.txt +1 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_b4fff7ba577a27e2c997+2c2d707e/model.hlo_module.pb +3 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_b4fff7ba577a27e2c997+2c2d707e/model.neff +3 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_f8b7d65c4a40ccde8769+2c2d707e/compile_flags.txt +1 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_f8b7d65c4a40ccde8769+2c2d707e/model.hlo_module.pb +3 -0
- neuronxcc-2.14.213.0+013d129b/MODULE_f8b7d65c4a40ccde8769+2c2d707e/model.neff +3 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/2c73925e46a18751b152.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/7b3393a7c5fd984a34d9.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/c997564eba3d6777671f.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/gpt2/720f4c07d5f83b2ff9d4.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/1fb2d1d6f793fd1c9744.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/65314683f5d624301ac7.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/f32b4f1b111de9eb757b.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/1caf4d7e91344cb4edd4.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/a5d3d8806cd364c7a404.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/cc9c0a9006808dddfa27.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/17dbb731c960381a466f.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/770ede8d2256a26e483f.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/7e2a0c353c70de261074.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/mixtral/dacorvo/Mixtral-tiny/000f1688e11ccb68df60.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/mixtral/dacorvo/Mixtral-tiny/089a1107126e0bb3ee3d.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/mixtral/dacorvo/Mixtral-tiny/abfe2aa64af6a13ba7f8.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/09dbce9d84c145ef699b.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/10b079960341f41cb11b.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/ddbc5f019bf5542a2be6.json +1 -0
- neuronxcc-2.14.227.0+2d4f85be/MODULE_1ddfe7bb03e6b5141bca+2c2d707e/compile_flags.txt +1 -0
- neuronxcc-2.14.227.0+2d4f85be/MODULE_1ddfe7bb03e6b5141bca+2c2d707e/model.hlo_module.pb +3 -0
- neuronxcc-2.14.227.0+2d4f85be/MODULE_1ddfe7bb03e6b5141bca+2c2d707e/model.neff +0 -0
- neuronxcc-2.14.227.0+2d4f85be/MODULE_1f3d02b2530b83c4abf7+2c2d707e/compile_flags.txt +1 -0
- neuronxcc-2.14.227.0+2d4f85be/MODULE_1f3d02b2530b83c4abf7+2c2d707e/model.hlo_module.pb +3 -0
- neuronxcc-2.14.227.0+2d4f85be/MODULE_1f3d02b2530b83c4abf7+2c2d707e/model.neff +0 -0
- neuronxcc-2.14.227.0+2d4f85be/MODULE_23ced3c113e97afdb062+2c2d707e/compile_flags.txt +1 -0
- neuronxcc-2.14.227.0+2d4f85be/MODULE_23ced3c113e97afdb062+2c2d707e/model.hlo_module.pb +3 -0
.gitattributes
CHANGED
@@ -449,3 +449,11 @@ neuronxcc-2.14.213.0+013d129b/MODULE_ec7ce1cb59fb29fdc73b+2c2d707e/model.neff fi
|
|
449 |
neuronxcc-2.14.213.0+013d129b/MODULE_ed56de1578306228c0e9+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
450 |
neuronxcc-2.14.213.0+013d129b/MODULE_f3c3e815bc8fb3c7b0a3+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
451 |
neuronxcc-2.14.213.0+013d129b/MODULE_f5819d12264aac7692e2+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
449 |
neuronxcc-2.14.213.0+013d129b/MODULE_ed56de1578306228c0e9+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
450 |
neuronxcc-2.14.213.0+013d129b/MODULE_f3c3e815bc8fb3c7b0a3+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
451 |
neuronxcc-2.14.213.0+013d129b/MODULE_f5819d12264aac7692e2+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
452 |
+
neuronxcc-2.14.213.0+013d129b/MODULE_2354da4d82d493971ffa+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
453 |
+
neuronxcc-2.14.213.0+013d129b/MODULE_2f1b54731e540b7a1e39+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
454 |
+
neuronxcc-2.14.213.0+013d129b/MODULE_45c97e1b335e74b7fcf4+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
455 |
+
neuronxcc-2.14.213.0+013d129b/MODULE_58ddfb26c2d4844ddc48+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
456 |
+
neuronxcc-2.14.213.0+013d129b/MODULE_b4fff7ba577a27e2c997+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
457 |
+
neuronxcc-2.14.213.0+013d129b/MODULE_f8b7d65c4a40ccde8769+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
458 |
+
neuronxcc-2.14.227.0+2d4f85be/MODULE_345116edd09879b7ebe4+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
459 |
+
neuronxcc-2.14.227.0+2d4f85be/MODULE_e1e4cdeee8e8e42ef9b0+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
neuronxcc-2.14.213.0+013d129b/MODULE_2354da4d82d493971ffa+2c2d707e/compile_flags.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
--model-type=transformer --auto-cast=none
|
neuronxcc-2.14.213.0+013d129b/MODULE_2354da4d82d493971ffa+2c2d707e/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:837cd2f5033c185d23c028462ab9a86eaef090aaded40d381bc48cbc029afa29
|
3 |
+
size 267317
|
neuronxcc-2.14.213.0+013d129b/MODULE_2354da4d82d493971ffa+2c2d707e/model.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:96f9eba5ac448af039548544ce186a4f94126b39df2d4d8af9b82a0234fe4257
|
3 |
+
size 43070464
|
neuronxcc-2.14.213.0+013d129b/MODULE_2f1b54731e540b7a1e39+2c2d707e/compile_flags.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
--model-type=transformer --auto-cast=none
|
neuronxcc-2.14.213.0+013d129b/MODULE_2f1b54731e540b7a1e39+2c2d707e/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:789b8d5f526a0d99c49c1d949ed4f43e7f04b8804c9ee56dbbf31d418a57c1fb
|
3 |
+
size 335033
|
neuronxcc-2.14.213.0+013d129b/MODULE_2f1b54731e540b7a1e39+2c2d707e/model.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a88dbe0f2142748627cd842e795697dcf0a91261dacc464a0a075c48334032ef
|
3 |
+
size 11377664
|
neuronxcc-2.14.213.0+013d129b/MODULE_45c97e1b335e74b7fcf4+2c2d707e/compile_flags.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
--model-type=transformer --auto-cast=none
|
neuronxcc-2.14.213.0+013d129b/MODULE_45c97e1b335e74b7fcf4+2c2d707e/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a151edfe288773ee715ca57fe960427ddd733aaaead9d996574e348e7cd05c99
|
3 |
+
size 233333
|
neuronxcc-2.14.213.0+013d129b/MODULE_45c97e1b335e74b7fcf4+2c2d707e/model.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26cca47385427ec5dee6bc3bf855a6028018d3fffe6aa798232792800ae809a5
|
3 |
+
size 3902464
|
neuronxcc-2.14.213.0+013d129b/MODULE_58ddfb26c2d4844ddc48+2c2d707e/compile_flags.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
--model-type=transformer --auto-cast=none
|
neuronxcc-2.14.213.0+013d129b/MODULE_58ddfb26c2d4844ddc48+2c2d707e/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fa8889a59c70ee5fca60978dfe211f2480bf757ddd9be1b90ca057d2acc733e8
|
3 |
+
size 450118
|
neuronxcc-2.14.213.0+013d129b/MODULE_58ddfb26c2d4844ddc48+2c2d707e/model.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c66c47ac9860811718950f9071b6deabe113365eb8ca0f005509a458ba0af4be
|
3 |
+
size 10271744
|
neuronxcc-2.14.213.0+013d129b/MODULE_94243e60d465b93319ce+2c2d707e/compile_flags.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
--model-type=transformer --auto-cast=none
|
neuronxcc-2.14.213.0+013d129b/MODULE_94243e60d465b93319ce+2c2d707e/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c66620514d2824f0de22b69eacdb9fca080e1da326ae6505ed7802f753c8822c
|
3 |
+
size 274512
|
neuronxcc-2.14.213.0+013d129b/MODULE_a6dc5c5ed3cdf7cd3eca+2c2d707e/compile_flags.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
--model-type=transformer --auto-cast=none
|
neuronxcc-2.14.213.0+013d129b/MODULE_a6dc5c5ed3cdf7cd3eca+2c2d707e/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c5bec7c92165be15bb09a54ef6810f447d643aba9591dc463c82606b66ee5071
|
3 |
+
size 358850
|
neuronxcc-2.14.213.0+013d129b/MODULE_b4fff7ba577a27e2c997+2c2d707e/compile_flags.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
--model-type=transformer --auto-cast=none
|
neuronxcc-2.14.213.0+013d129b/MODULE_b4fff7ba577a27e2c997+2c2d707e/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7da21af3749343c5c27f63d72ba37a29fc4f91e9d3d1ddb4117e0877f8ed73aa
|
3 |
+
size 450118
|
neuronxcc-2.14.213.0+013d129b/MODULE_b4fff7ba577a27e2c997+2c2d707e/model.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:27fa4be5edcdf24b9acaffa94ec8dbeb46c351701a7984094197dea1db8143ed
|
3 |
+
size 10271744
|
neuronxcc-2.14.213.0+013d129b/MODULE_f8b7d65c4a40ccde8769+2c2d707e/compile_flags.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
--model-type=transformer --auto-cast=none
|
neuronxcc-2.14.213.0+013d129b/MODULE_f8b7d65c4a40ccde8769+2c2d707e/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d18ec0ed5e448caefe7bb4300507568a1bc9720260acee751efb7b7ac7982b60
|
3 |
+
size 191491
|
neuronxcc-2.14.213.0+013d129b/MODULE_f8b7d65c4a40ccde8769+2c2d707e/model.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:063d6d223278c2e6774573a6415bfea68fe03c8b285b6d8fdb285c20581b5521
|
3 |
+
size 13620224
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/2c73925e46a18751b152.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/7b3393a7c5fd984a34d9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/c997564eba3d6777671f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/gpt2/720f4c07d5f83b2ff9d4.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/1fb2d1d6f793fd1c9744.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/65314683f5d624301ac7.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/f32b4f1b111de9eb757b.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/1caf4d7e91344cb4edd4.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/a5d3d8806cd364c7a404.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/cc9c0a9006808dddfa27.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/17dbb731c960381a466f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/770ede8d2256a26e483f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/7e2a0c353c70de261074.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/mixtral/dacorvo/Mixtral-tiny/000f1688e11ccb68df60.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/mixtral/dacorvo/Mixtral-tiny/089a1107126e0bb3ee3d.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/mixtral/dacorvo/Mixtral-tiny/abfe2aa64af6a13ba7f8.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/09dbce9d84c145ef699b.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/10b079960341f41cb11b.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/ddbc5f019bf5542a2be6.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.14.227.0+2d4f85be/MODULE_1ddfe7bb03e6b5141bca+2c2d707e/compile_flags.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
--model-type=transformer --auto-cast=none
|
neuronxcc-2.14.227.0+2d4f85be/MODULE_1ddfe7bb03e6b5141bca+2c2d707e/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0beb3bc81714d1f70e77131f347a87578e245d93457b5aa875e18761c434e459
|
3 |
+
size 1900
|
neuronxcc-2.14.227.0+2d4f85be/MODULE_1ddfe7bb03e6b5141bca+2c2d707e/model.neff
ADDED
Binary file (42 kB). View file
|
|
neuronxcc-2.14.227.0+2d4f85be/MODULE_1f3d02b2530b83c4abf7+2c2d707e/compile_flags.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
--model-type=transformer --auto-cast=none
|
neuronxcc-2.14.227.0+2d4f85be/MODULE_1f3d02b2530b83c4abf7+2c2d707e/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c016cb9348980e5ceadfeb94168eab02c144bb5693c61751c2d3367d09ef5b5f
|
3 |
+
size 57433
|
neuronxcc-2.14.227.0+2d4f85be/MODULE_1f3d02b2530b83c4abf7+2c2d707e/model.neff
ADDED
Binary file (155 kB). View file
|
|
neuronxcc-2.14.227.0+2d4f85be/MODULE_23ced3c113e97afdb062+2c2d707e/compile_flags.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
--model-type=transformer --auto-cast=none
|
neuronxcc-2.14.227.0+2d4f85be/MODULE_23ced3c113e97afdb062+2c2d707e/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:76ee8fcd669c0344010019e40d232b2beb49ae91140f62839d3464f9d6c4bd3a
|
3 |
+
size 60871
|