samir-souza commited on
Commit
7a4b0b4
·
verified ·
1 Parent(s): 57e08fd

Synchronizing local compiler cache.

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +6 -0
  2. neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/training/llama/meta-llama/Llama-3.2-1B/9276854d78c440ee3d11.json +1 -0
  3. neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/training/llama/meta-llama/Llama-3.2-1B/eccc0ed3e0e331d14aea.json +1 -0
  4. neuronxcc-2.15.128.0+56dc5a86/MODULE_10812541582330903242+ebf00a7f/compile_flags.json +1 -0
  5. neuronxcc-2.15.128.0+56dc5a86/MODULE_10812541582330903242+ebf00a7f/model.done +0 -0
  6. neuronxcc-2.15.128.0+56dc5a86/MODULE_10812541582330903242+ebf00a7f/model.hlo_module.pb +3 -0
  7. neuronxcc-2.15.128.0+56dc5a86/MODULE_10812541582330903242+ebf00a7f/model.neff +3 -0
  8. neuronxcc-2.15.128.0+56dc5a86/MODULE_11043275446542233300+ebf00a7f/compile_flags.json +1 -0
  9. neuronxcc-2.15.128.0+56dc5a86/MODULE_11043275446542233300+ebf00a7f/model.done +0 -0
  10. neuronxcc-2.15.128.0+56dc5a86/MODULE_11043275446542233300+ebf00a7f/model.hlo_module.pb +3 -0
  11. neuronxcc-2.15.128.0+56dc5a86/MODULE_11043275446542233300+ebf00a7f/model.neff +0 -0
  12. neuronxcc-2.15.128.0+56dc5a86/MODULE_11941768716833979484+ebf00a7f/compile_flags.json +1 -0
  13. neuronxcc-2.15.128.0+56dc5a86/MODULE_11941768716833979484+ebf00a7f/model.done +0 -0
  14. neuronxcc-2.15.128.0+56dc5a86/MODULE_11941768716833979484+ebf00a7f/model.hlo_module.pb +3 -0
  15. neuronxcc-2.15.128.0+56dc5a86/MODULE_11941768716833979484+ebf00a7f/model.neff +0 -0
  16. neuronxcc-2.15.128.0+56dc5a86/MODULE_13855921216246581326+ebf00a7f/compile_flags.json +1 -0
  17. neuronxcc-2.15.128.0+56dc5a86/MODULE_13855921216246581326+ebf00a7f/model.done +0 -0
  18. neuronxcc-2.15.128.0+56dc5a86/MODULE_13855921216246581326+ebf00a7f/model.hlo_module.pb +3 -0
  19. neuronxcc-2.15.128.0+56dc5a86/MODULE_13855921216246581326+ebf00a7f/model.neff +3 -0
  20. neuronxcc-2.15.128.0+56dc5a86/MODULE_16413446298350855817+ebf00a7f/compile_flags.json +1 -0
  21. neuronxcc-2.15.128.0+56dc5a86/MODULE_16413446298350855817+ebf00a7f/model.done +0 -0
  22. neuronxcc-2.15.128.0+56dc5a86/MODULE_16413446298350855817+ebf00a7f/model.hlo_module.pb +3 -0
  23. neuronxcc-2.15.128.0+56dc5a86/MODULE_16413446298350855817+ebf00a7f/model.neff +0 -0
  24. neuronxcc-2.15.128.0+56dc5a86/MODULE_18104239624254157043+ebf00a7f/compile_flags.json +1 -0
  25. neuronxcc-2.15.128.0+56dc5a86/MODULE_18104239624254157043+ebf00a7f/model.done +0 -0
  26. neuronxcc-2.15.128.0+56dc5a86/MODULE_18104239624254157043+ebf00a7f/model.hlo_module.pb +3 -0
  27. neuronxcc-2.15.128.0+56dc5a86/MODULE_18104239624254157043+ebf00a7f/model.neff +0 -0
  28. neuronxcc-2.15.128.0+56dc5a86/MODULE_2139530822776559794+ebf00a7f/compile_flags.json +1 -0
  29. neuronxcc-2.15.128.0+56dc5a86/MODULE_2139530822776559794+ebf00a7f/model.done +0 -0
  30. neuronxcc-2.15.128.0+56dc5a86/MODULE_2139530822776559794+ebf00a7f/model.hlo_module.pb +3 -0
  31. neuronxcc-2.15.128.0+56dc5a86/MODULE_2139530822776559794+ebf00a7f/model.neff +0 -0
  32. neuronxcc-2.15.128.0+56dc5a86/MODULE_3198840761297369981+ebf00a7f/compile_flags.json +1 -0
  33. neuronxcc-2.15.128.0+56dc5a86/MODULE_3198840761297369981+ebf00a7f/model.done +0 -0
  34. neuronxcc-2.15.128.0+56dc5a86/MODULE_3198840761297369981+ebf00a7f/model.hlo_module.pb +3 -0
  35. neuronxcc-2.15.128.0+56dc5a86/MODULE_3198840761297369981+ebf00a7f/model.neff +3 -0
  36. neuronxcc-2.15.128.0+56dc5a86/MODULE_4161043282471674704+ebf00a7f/compile_flags.json +1 -0
  37. neuronxcc-2.15.128.0+56dc5a86/MODULE_4161043282471674704+ebf00a7f/model.done +0 -0
  38. neuronxcc-2.15.128.0+56dc5a86/MODULE_4161043282471674704+ebf00a7f/model.hlo_module.pb +3 -0
  39. neuronxcc-2.15.128.0+56dc5a86/MODULE_4161043282471674704+ebf00a7f/model.neff +0 -0
  40. neuronxcc-2.15.128.0+56dc5a86/MODULE_4938805684643787240+ebf00a7f/compile_flags.json +1 -0
  41. neuronxcc-2.15.128.0+56dc5a86/MODULE_4938805684643787240+ebf00a7f/model.done +0 -0
  42. neuronxcc-2.15.128.0+56dc5a86/MODULE_4938805684643787240+ebf00a7f/model.hlo_module.pb +3 -0
  43. neuronxcc-2.15.128.0+56dc5a86/MODULE_4938805684643787240+ebf00a7f/model.neff +0 -0
  44. neuronxcc-2.15.128.0+56dc5a86/MODULE_5086399037193202461+ebf00a7f/compile_flags.json +1 -0
  45. neuronxcc-2.15.128.0+56dc5a86/MODULE_5086399037193202461+ebf00a7f/model.done +0 -0
  46. neuronxcc-2.15.128.0+56dc5a86/MODULE_5086399037193202461+ebf00a7f/model.hlo_module.pb +3 -0
  47. neuronxcc-2.15.128.0+56dc5a86/MODULE_5086399037193202461+ebf00a7f/model.neff +3 -0
  48. neuronxcc-2.15.128.0+56dc5a86/MODULE_5104869544532712658+ebf00a7f/compile_flags.json +1 -0
  49. neuronxcc-2.15.128.0+56dc5a86/MODULE_5104869544532712658+ebf00a7f/model.done +0 -0
  50. neuronxcc-2.15.128.0+56dc5a86/MODULE_5104869544532712658+ebf00a7f/model.hlo_module.pb +3 -0
.gitattributes CHANGED
@@ -5705,3 +5705,9 @@ neuronxcc-2.15.128.0+56dc5a86/MODULE_5352243160435722256+0662544f/model.neff fil
5705
  neuronxcc-2.15.128.0+56dc5a86/MODULE_9004138607823884889+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
5706
  neuronxcc-2.15.128.0+56dc5a86/MODULE_355980366824992736+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
5707
  neuronxcc-2.15.128.0+56dc5a86/MODULE_6220973826920381074+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
5705
  neuronxcc-2.15.128.0+56dc5a86/MODULE_9004138607823884889+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
5706
  neuronxcc-2.15.128.0+56dc5a86/MODULE_355980366824992736+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
5707
  neuronxcc-2.15.128.0+56dc5a86/MODULE_6220973826920381074+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
5708
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_10812541582330903242+ebf00a7f/model.neff filter=lfs diff=lfs merge=lfs -text
5709
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_13855921216246581326+ebf00a7f/model.neff filter=lfs diff=lfs merge=lfs -text
5710
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_3198840761297369981+ebf00a7f/model.neff filter=lfs diff=lfs merge=lfs -text
5711
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_5086399037193202461+ebf00a7f/model.neff filter=lfs diff=lfs merge=lfs -text
5712
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_5104869544532712658+ebf00a7f/model.neff filter=lfs diff=lfs merge=lfs -text
5713
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_6014407322135731921+ebf00a7f/model.neff filter=lfs diff=lfs merge=lfs -text
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/training/llama/meta-llama/Llama-3.2-1B/9276854d78c440ee3d11.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128001, "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"compiler_version": "2.15.128.0+56dc5a86", "input_specs": {"chosen": ["To prove whether f is a constant function or not, we will use Liouville's theorem. Liouville's theorem states that any bounded entire function (i.e., a function that is holomorphic on the entire complex plane and is bounded) must be constant.\n\nLet's assume that f(z) is not a constant function. Then there exist z1 and z2 in the complex plane C such that f(z1) \u2260 f(z2).\n\nNow, we need to find a bound on |f(z)|. We can start by considering the expression M = max{|f(z1)|, |f(z2)|}. By definition of maxima, |f(z)| \u2264 M for all z in C.\n\nNext, we will find a bounded entire function g(z) that is not constant. Let's define a function g(z) as g(z) = f(z) - 1/f(z2). Since f(z) is not constant, f(z) is not equal to zero anywhere in C. Thus, we can create a new function g(z) that takes into consideration the properties of f(z).\n\nNow let's analyze g(z). Notice that |g(z)| = |f(z) - 1/f(z2)|. We can bound |g(z)| as follows: \n\n1. If |f(z)| \u2265 |f(z2)|, then |g(z)| = |f(z) - 1/f(z2)| \u2264 |f(z)| + 1/|f(z2)| \u2264 M + 1/|f(z2)|, which is a constant since M is a constant and |f(z2)| is also a constant.\n\n2. If |f(z)| < |f(z2)|, then |g(z)| = |f(z) - 1/f(z2)| \u2264 |f(z)| + 1/|f(z2)| \u2264 |f(z2)| + 1/|f(z2)| \u2264 M, which is a constant since |f(z2)| is a constant.\n\nIn either case, we find that g(z) is a bounded entire function. Therefore, by Liouville's theorem, g(z) must be a constant function, which contradicts our assumption that f(z) is not constant.\n\nHence, our assumption that f(z) is not constant was incorrect. Therefore, f(z) must be a constant function."], "chosen_attention_mask": [1, 128], "chosen_input_ids": [1, 128], "chosen_labels": [1, 128], "prompt": ["Let f be an entire function such that |f(z)| $\\geq$ 1 for every z in C. Is f is a constant function?"], "prompt_attention_mask": [1, 128], "prompt_input_ids": [1, 128], "question": ["Let f be an entire function such that |f(z)| $\\geq$ 1 for every z in C. Is f is a constant function?"], "rejected": ["No, the entire function f cannot be a constant function because its modulus or absolute value on the entire complex plane, denoted by |f(z)|, is greater than or equal to 1. This means that f takes values outside the one-point set {c}, where c is any complex number, which is a necessary condition for a function to be a constant. Therefore, f must take at least two distinct values and hence is not a constant function."], "rejected_attention_mask": [1, 128], "rejected_input_ids": [1, 128], "rejected_labels": [1, 128], "source": ["TheoremQA"], "system": [null]}, "model_class": "LlamaForCausalLM", "num_neuron_cores_per_node": 2, "pipeline_parallel_size": 1, "precision": "bfloat16", "tensor_parallel_size": 1, "training": true}, "num_attention_heads": 32, "num_hidden_layers": 16, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 32.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/training/llama/meta-llama/Llama-3.2-1B/eccc0ed3e0e331d14aea.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128001, "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"compiler_version": "2.15.128.0+56dc5a86", "input_specs": {"chosen": ["This statement demonstrates the logical fallacy known as 'hasty generalization.' The speaker draws a faulty conclusion about vegetarians' impact on the environment based on a single, isolated example of a vegetarian who drives a gas-guzzling car. Anecdotal evidence like this cannot be used to make general claims about a larger population."], "chosen_attention_mask": [1, 128], "chosen_input_ids": [1, 128], "chosen_labels": [1, 128], "prompt": ["Identify the logical fallacy in this statement: 'Just because you're a vegetarian doesn't mean you're saving the environment. My friend's cousin is a vegetarian, and she drives a gas-guzzling car.'"], "prompt_attention_mask": [1, 128], "prompt_input_ids": [1, 128], "question": ["Identify the logical fallacy in this statement: 'Just because you're a vegetarian doesn't mean you're saving the environment. My friend's cousin is a vegetarian, and she drives a gas-guzzling car.'"], "rejected": ["This statement demonstrates the logical fallacy known as 'hasty generalization.' The speaker draws a faulty conclusion about vegetarians' impact on the environment based on a single, isolated example of a vegetarian who drives a gas-guzzling car. Anecdotal evidence like this cannot be used to make general claims about a larger population."], "rejected_attention_mask": [1, 128], "rejected_input_ids": [1, 128], "rejected_labels": [1, 128], "source": ["General-Instruct"], "system": [null]}, "model_class": "LlamaForCausalLM", "num_neuron_cores_per_node": 2, "pipeline_parallel_size": 1, "precision": "bfloat16", "tensor_parallel_size": 1, "training": true}, "num_attention_heads": 32, "num_hidden_layers": 16, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 32.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
neuronxcc-2.15.128.0+56dc5a86/MODULE_10812541582330903242+ebf00a7f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--distribution-strategy=llm-training", "--enable-saturate-infinity"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_10812541582330903242+ebf00a7f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_10812541582330903242+ebf00a7f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27ae554b07ad6e1d74e254f7702be9159e494cd02f580f35c53bef165ecf4ca6
3
+ size 170034
neuronxcc-2.15.128.0+56dc5a86/MODULE_10812541582330903242+ebf00a7f/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:575a11a6703fd6822a7f5b40540a7cc0745afca51b44d6d23becce65ae2f9adb
3
+ size 5929984
neuronxcc-2.15.128.0+56dc5a86/MODULE_11043275446542233300+ebf00a7f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--distribution-strategy=llm-training", "--enable-saturate-infinity"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_11043275446542233300+ebf00a7f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_11043275446542233300+ebf00a7f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4adf9ac4b0a832bf44afccd840bd91653159bdddd52124105cd62c8b3010b638
3
+ size 34493
neuronxcc-2.15.128.0+56dc5a86/MODULE_11043275446542233300+ebf00a7f/model.neff ADDED
Binary file (42 kB). View file
 
neuronxcc-2.15.128.0+56dc5a86/MODULE_11941768716833979484+ebf00a7f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--distribution-strategy=llm-training", "--enable-saturate-infinity"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_11941768716833979484+ebf00a7f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_11941768716833979484+ebf00a7f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3d340c40c070ca0b53599120118abf0ba2367268afe80c2b61c6e7df401a0d1
3
+ size 4477
neuronxcc-2.15.128.0+56dc5a86/MODULE_11941768716833979484+ebf00a7f/model.neff ADDED
Binary file (114 kB). View file
 
neuronxcc-2.15.128.0+56dc5a86/MODULE_13855921216246581326+ebf00a7f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--distribution-strategy=llm-training", "--enable-saturate-infinity"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_13855921216246581326+ebf00a7f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_13855921216246581326+ebf00a7f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:803893f3e2e655073be8d366d41b860af098c1a1a79a0f7ff14509f0842db439
3
+ size 673120
neuronxcc-2.15.128.0+56dc5a86/MODULE_13855921216246581326+ebf00a7f/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7e0985e8c8cf8b6289c87a507b478393786f13d61ddbcd4dc9d9d334ee05742
3
+ size 5530624
neuronxcc-2.15.128.0+56dc5a86/MODULE_16413446298350855817+ebf00a7f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--distribution-strategy=llm-training", "--enable-saturate-infinity"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_16413446298350855817+ebf00a7f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_16413446298350855817+ebf00a7f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4639515d756144f0ce4448cf5ecb6c239fc69a04c0ae410abeb7c32933c41580
3
+ size 1607
neuronxcc-2.15.128.0+56dc5a86/MODULE_16413446298350855817+ebf00a7f/model.neff ADDED
Binary file (21.5 kB). View file
 
neuronxcc-2.15.128.0+56dc5a86/MODULE_18104239624254157043+ebf00a7f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--distribution-strategy=llm-training", "--enable-saturate-infinity"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_18104239624254157043+ebf00a7f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_18104239624254157043+ebf00a7f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcb3816b08ace82ab5cebe5ef41490a42f4715fb39308b4360c2b9ef14c01824
3
+ size 1845
neuronxcc-2.15.128.0+56dc5a86/MODULE_18104239624254157043+ebf00a7f/model.neff ADDED
Binary file (31.7 kB). View file
 
neuronxcc-2.15.128.0+56dc5a86/MODULE_2139530822776559794+ebf00a7f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--distribution-strategy=llm-training", "--enable-saturate-infinity"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_2139530822776559794+ebf00a7f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_2139530822776559794+ebf00a7f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc4da2257799aad79dbd6e3b21b7d85a46ae8b9f195df50fa028d6a8a1072de2
3
+ size 1424
neuronxcc-2.15.128.0+56dc5a86/MODULE_2139530822776559794+ebf00a7f/model.neff ADDED
Binary file (21.5 kB). View file
 
neuronxcc-2.15.128.0+56dc5a86/MODULE_3198840761297369981+ebf00a7f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--distribution-strategy=llm-training", "--enable-saturate-infinity"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_3198840761297369981+ebf00a7f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_3198840761297369981+ebf00a7f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f399dc293652ea0666dbd8f099bb719b8c03d405669577578501b1fa65e84f2
3
+ size 347095
neuronxcc-2.15.128.0+56dc5a86/MODULE_3198840761297369981+ebf00a7f/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:309eb79d45eba1b2642230cfe2ee746f156f32926377022f1553ee6b8b71861e
3
+ size 12329984
neuronxcc-2.15.128.0+56dc5a86/MODULE_4161043282471674704+ebf00a7f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--distribution-strategy=llm-training", "--enable-saturate-infinity"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_4161043282471674704+ebf00a7f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_4161043282471674704+ebf00a7f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d749034fe137babe440d77281eaf18d107f8daec79912d2ebcfac961115eae93
3
+ size 4936
neuronxcc-2.15.128.0+56dc5a86/MODULE_4161043282471674704+ebf00a7f/model.neff ADDED
Binary file (114 kB). View file
 
neuronxcc-2.15.128.0+56dc5a86/MODULE_4938805684643787240+ebf00a7f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--distribution-strategy=llm-training", "--enable-saturate-infinity"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_4938805684643787240+ebf00a7f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_4938805684643787240+ebf00a7f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b91e3814118ae5c4b3dd56b8efa4b98b9b78d8497f1112478beb9d97cd6147ef
3
+ size 1424
neuronxcc-2.15.128.0+56dc5a86/MODULE_4938805684643787240+ebf00a7f/model.neff ADDED
Binary file (21.5 kB). View file
 
neuronxcc-2.15.128.0+56dc5a86/MODULE_5086399037193202461+ebf00a7f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--distribution-strategy=llm-training", "--enable-saturate-infinity"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_5086399037193202461+ebf00a7f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_5086399037193202461+ebf00a7f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ba9c769c9d465216783b627cfbc1d4e9b0b60067d6471c25f3c8230c0938834
3
+ size 624360
neuronxcc-2.15.128.0+56dc5a86/MODULE_5086399037193202461+ebf00a7f/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7145f6d7a9d5b4fbdc69862debc98d775a1c9c4688ddc555687dd824f3df78b3
3
+ size 4527104
neuronxcc-2.15.128.0+56dc5a86/MODULE_5104869544532712658+ebf00a7f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--distribution-strategy=llm-training", "--enable-saturate-infinity"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_5104869544532712658+ebf00a7f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_5104869544532712658+ebf00a7f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:815d5b425f7f7b5fc456218c238c910bf929dad95cc0857faa21e81fd7187af7
3
+ size 242194