Text Generation
Transformers
llama
Inference Endpoints
mpasila commited on
Commit
3235f1c
1 Parent(s): f0a0feb

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. README.md +16 -14
  2. cal_data.safetensors +3 -0
  3. config.json +28 -39
  4. generation_config.json +1 -1
  5. hidden_states.safetensors +3 -0
  6. job_new.json +0 -0
  7. measurement.json +0 -0
  8. out_tensor/lm_head.safetensors +3 -0
  9. out_tensor/model.layers.0.mlp.down_proj.safetensors +3 -0
  10. out_tensor/model.layers.0.mlp.gate_proj.safetensors +3 -0
  11. out_tensor/model.layers.0.mlp.up_proj.safetensors +3 -0
  12. out_tensor/model.layers.0.self_attn.k_proj.safetensors +3 -0
  13. out_tensor/model.layers.0.self_attn.o_proj.safetensors +3 -0
  14. out_tensor/model.layers.0.self_attn.q_proj.safetensors +3 -0
  15. out_tensor/model.layers.0.self_attn.v_proj.safetensors +3 -0
  16. out_tensor/model.layers.1.mlp.down_proj.safetensors +3 -0
  17. out_tensor/model.layers.1.mlp.gate_proj.safetensors +3 -0
  18. out_tensor/model.layers.1.mlp.up_proj.safetensors +3 -0
  19. out_tensor/model.layers.1.self_attn.k_proj.safetensors +3 -0
  20. out_tensor/model.layers.1.self_attn.o_proj.safetensors +3 -0
  21. out_tensor/model.layers.1.self_attn.q_proj.safetensors +3 -0
  22. out_tensor/model.layers.1.self_attn.v_proj.safetensors +3 -0
  23. out_tensor/model.layers.10.mlp.down_proj.safetensors +3 -0
  24. out_tensor/model.layers.10.mlp.gate_proj.safetensors +3 -0
  25. out_tensor/model.layers.10.mlp.up_proj.safetensors +3 -0
  26. out_tensor/model.layers.10.self_attn.k_proj.safetensors +3 -0
  27. out_tensor/model.layers.10.self_attn.o_proj.safetensors +3 -0
  28. out_tensor/model.layers.10.self_attn.q_proj.safetensors +3 -0
  29. out_tensor/model.layers.10.self_attn.v_proj.safetensors +3 -0
  30. out_tensor/model.layers.11.mlp.down_proj.safetensors +3 -0
  31. out_tensor/model.layers.11.mlp.gate_proj.safetensors +3 -0
  32. out_tensor/model.layers.11.mlp.up_proj.safetensors +3 -0
  33. out_tensor/model.layers.11.self_attn.k_proj.safetensors +3 -0
  34. out_tensor/model.layers.11.self_attn.o_proj.safetensors +3 -0
  35. out_tensor/model.layers.11.self_attn.q_proj.safetensors +3 -0
  36. out_tensor/model.layers.11.self_attn.v_proj.safetensors +3 -0
  37. out_tensor/model.layers.12.mlp.down_proj.safetensors +3 -0
  38. out_tensor/model.layers.12.mlp.gate_proj.safetensors +3 -0
  39. out_tensor/model.layers.12.mlp.up_proj.safetensors +3 -0
  40. out_tensor/model.layers.12.self_attn.k_proj.safetensors +3 -0
  41. out_tensor/model.layers.12.self_attn.o_proj.safetensors +3 -0
  42. out_tensor/model.layers.12.self_attn.q_proj.safetensors +3 -0
  43. out_tensor/model.layers.12.self_attn.v_proj.safetensors +3 -0
  44. out_tensor/model.layers.13.mlp.down_proj.safetensors +3 -0
  45. out_tensor/model.layers.13.mlp.gate_proj.safetensors +3 -0
  46. out_tensor/model.layers.13.mlp.up_proj.safetensors +3 -0
  47. out_tensor/model.layers.13.self_attn.k_proj.safetensors +3 -0
  48. out_tensor/model.layers.13.self_attn.o_proj.safetensors +3 -0
  49. out_tensor/model.layers.13.self_attn.q_proj.safetensors +3 -0
  50. out_tensor/model.layers.13.self_attn.v_proj.safetensors +3 -0
README.md CHANGED
@@ -12,21 +12,13 @@ language:
12
  - 'no'
13
  - nn
14
  - is
15
- base_model: LumiOpen/Viking-7B
16
  ---
17
- NOTE THIS IS BASED ON THE 1000B VERSION! I will update this soon.
18
-
19
- This is an ExLlamaV2 quantized model in 4bpw of [LumiOpen/Viking-7B](https://huggingface.co/LumiOpen/Viking-7B) using the default calibration dataset.
20
-
21
- # Original Model card:
22
 
23
  # Viking 7B
24
 
25
- **NOTE: We are aware of an incompatibility with HF transformers that impacts finetuning and are working to correct it.**
26
-
27
- _**NOTE:** This is a **research checkpoint** of a model for which **training has not been completed.** It is being provided in its current state for research and testing purposes. **Care should be taken when using the outputs of the model.** Once pretraining has completed we intend to release additional instruction-tuned and chat-tuned varieties._
28
-
29
- Viking 7B is a 7B parameter decoder-only transformer pretrained on Finnish, English, Swedish, Danish, Norwegian, Icelandic and code. It is being trained on 2 trillion tokens (1 trillion as of this release). Viking 7B is a fully open source model and is made available under the Apache 2.0 License.
30
 
31
  Viking was created in a collaboration between the [TurkuNLP group](https://turkunlp.org/) of the University of Turku, [SiloGen](https://www.silo.ai/silogen) from [Silo AI](https://www.silo.ai/),and [High Performance Language Technologies](https://hplt-project.org/) (HPLT). Training was conducted on the [LUMI supercomputer](https://www.lumi-supercomputer.eu/), using compute resources generously provided by [CSC](https://csc.fi/) - IT Center for Science, Finland.
32
 
@@ -45,7 +37,7 @@ Viking is the second set of models released by LumiOpen and is available at
45
 
46
 
47
  ## Model Overview
48
- _**NOTE:** In addition to being an early research release, Viking is a base model which needs further fine tuning for most use cases._
49
 
50
  Viking is a generative pretrained transformer using a LLaMA-like GPT architecture, and makes use of rotary positional embeddings and flash attention.
51
 
@@ -103,14 +95,24 @@ Training Checkpoints are available as branches in the repository. Checkpoints w
103
  * [800B](https://huggingface.co/LumiOpen/Viking-7B/tree/800B)
104
  * [900B](https://huggingface.co/LumiOpen/Viking-7B/tree/900B)
105
  * [1000B](https://huggingface.co/LumiOpen/Viking-7B/tree/1000B)
 
 
 
 
 
 
 
 
 
 
106
 
107
  The transformers library allows you to load a checkpoint from a branch as follows:
108
 
109
  ```python
110
- branch = "200B"
111
  model = transformers.AutoModelForCausalLM.from_pretrained(
112
  "LumiOpen/Viking-7B",
113
- torch_dtype=torch.bfloat16,
114
  revision=branch,
115
  )
116
  ```
 
12
  - 'no'
13
  - nn
14
  - is
 
15
  ---
 
 
 
 
 
16
 
17
  # Viking 7B
18
 
19
+ Viking 7B is a 7B parameter decoder-only transformer pretrained on Finnish,
20
+ English, Swedish, Danish, Norwegian, Icelandic and code. It has been trained
21
+ on 2 trillion tokens. Viking 7B is a fully open source model and is made available under the Apache 2.0 License.
 
 
22
 
23
  Viking was created in a collaboration between the [TurkuNLP group](https://turkunlp.org/) of the University of Turku, [SiloGen](https://www.silo.ai/silogen) from [Silo AI](https://www.silo.ai/),and [High Performance Language Technologies](https://hplt-project.org/) (HPLT). Training was conducted on the [LUMI supercomputer](https://www.lumi-supercomputer.eu/), using compute resources generously provided by [CSC](https://csc.fi/) - IT Center for Science, Finland.
24
 
 
37
 
38
 
39
  ## Model Overview
40
+ _**NOTE:** Viking is a base model which needs further fine tuning for most use cases._
41
 
42
  Viking is a generative pretrained transformer using a LLaMA-like GPT architecture, and makes use of rotary positional embeddings and flash attention.
43
 
 
95
  * [800B](https://huggingface.co/LumiOpen/Viking-7B/tree/800B)
96
  * [900B](https://huggingface.co/LumiOpen/Viking-7B/tree/900B)
97
  * [1000B](https://huggingface.co/LumiOpen/Viking-7B/tree/1000B)
98
+ * [1100B](https://huggingface.co/LumiOpen/Viking-7B/tree/1100B)
99
+ * [1200B](https://huggingface.co/LumiOpen/Viking-7B/tree/1200B)
100
+ * [1300B](https://huggingface.co/LumiOpen/Viking-7B/tree/1300B)
101
+ * [1400B](https://huggingface.co/LumiOpen/Viking-7B/tree/1400B)
102
+ * [1500B](https://huggingface.co/LumiOpen/Viking-7B/tree/1500B)
103
+ * [1600B](https://huggingface.co/LumiOpen/Viking-7B/tree/1600B)
104
+ * [1700B](https://huggingface.co/LumiOpen/Viking-7B/tree/1700B)
105
+ * [1800B](https://huggingface.co/LumiOpen/Viking-7B/tree/1800B)
106
+ * [1900B](https://huggingface.co/LumiOpen/Viking-7B/tree/1900B)
107
+ * [2000B](https://huggingface.co/LumiOpen/Viking-7B/tree/2000B)
108
 
109
  The transformers library allows you to load a checkpoint from a branch as follows:
110
 
111
  ```python
112
+ branch = "2000B"
113
  model = transformers.AutoModelForCausalLM.from_pretrained(
114
  "LumiOpen/Viking-7B",
115
+ torch_dtype=torch.bfloat16 if torch.cuda.is_bf16_supported() else torch.float16,
116
  revision=branch,
117
  )
118
  ```
cal_data.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:040e73fd812b7abb24844b6a5bd261114780b01ea3c1cf908f009e3734e35c2c
3
+ size 1884248
config.json CHANGED
@@ -1,40 +1,29 @@
1
  {
2
- "_name_or_path": "/scratch/project_462000319/general-tools/viking_v2_checkpoints/viking_v2_7B_iter_0239000_bfloat16",
3
- "architectures": [
4
- "LlamaForCausalLM"
5
- ],
6
- "attention_bias": false,
7
- "attention_dropout": 0.0,
8
- "bos_token_id": 1,
9
- "eos_token_id": 2,
10
- "hidden_act": "silu",
11
- "hidden_size": 4096,
12
- "initializer_range": 0.02,
13
- "intermediate_size": 11008,
14
- "max_position_embeddings": 4096,
15
- "model_type": "llama",
16
- "num_attention_heads": 32,
17
- "num_hidden_layers": 32,
18
- "num_key_value_heads": 32,
19
- "pretraining_tp": 1,
20
- "rms_norm_eps": 1e-05,
21
- "rope_scaling": null,
22
- "rope_theta": 10000.0,
23
- "tie_word_embeddings": false,
24
- "torch_dtype": "bfloat16",
25
- "transformers_version": "4.37.2",
26
- "untie_embeddings_and_output_weights": true,
27
- "use_cache": true,
28
- "vocab_size": 131072,
29
- "quantization_config": {
30
- "quant_method": "exl2",
31
- "version": "0.0.17",
32
- "bits": 4.0,
33
- "head_bits": 6,
34
- "calibration": {
35
- "rows": 100,
36
- "length": 2048,
37
- "dataset": "(default)"
38
- }
39
- }
40
- }
 
1
  {
2
+ "_name_or_path": "/scratch/project_462000319/general-tools/viking_v2_checkpoints/viking_v2_7B_iter_0476837_bfloat16",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 4096,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 11008,
14
+ "max_position_embeddings": 4096,
15
+ "model_type": "llama",
16
+ "num_attention_heads": 32,
17
+ "num_hidden_layers": 32,
18
+ "num_key_value_heads": 32,
19
+ "pretraining_tp": 1,
20
+ "rms_norm_eps": 1e-05,
21
+ "rope_scaling": null,
22
+ "rope_theta": 10000.0,
23
+ "tie_word_embeddings": false,
24
+ "torch_dtype": "bfloat16",
25
+ "transformers_version": "4.40.0",
26
+ "untie_embeddings_and_output_weights": true,
27
+ "use_cache": true,
28
+ "vocab_size": 131072
29
+ }
 
 
 
 
 
 
 
 
 
 
 
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.37.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.40.0"
6
  }
hidden_states.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:401a35eac0cac36c6086067383005a154c330ed7ce6b4185d0a24be43056fe2f
3
+ size 1929389952
job_new.json CHANGED
The diff for this file is too large to render. See raw diff
 
measurement.json CHANGED
The diff for this file is too large to render. See raw diff
 
out_tensor/lm_head.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0298b05a794f4d85a43f1e76a2170d4c0e15833c86d8c1df133db2a712b4471e
3
+ size 425738872
out_tensor/model.layers.0.mlp.down_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90211f69276cd8e0a7fde44f40f425050db1ace7caf42c87e02f237aca409ffe
3
+ size 24563824
out_tensor/model.layers.0.mlp.gate_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b78a448d824f34eb79f324507c4aa1134ef6e116d72ea37225c5affa50d6855c
3
+ size 23442136
out_tensor/model.layers.0.mlp.up_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3b52d9a3b9a597101a46160f4d555b6671407d02e08fc3c3749e990bde3ed6e
3
+ size 24146632
out_tensor/model.layers.0.self_attn.k_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45f16f4dfa35cb9969ed3476624d9a900d7263edaacc87ec0b585bc906023e8c
3
+ size 8766368
out_tensor/model.layers.0.self_attn.o_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83500b7f6aeaef36abe4139bd6fbcdf1e3dc80ee2d9dc389b407288b4ff4cd1f
3
+ size 8766368
out_tensor/model.layers.0.self_attn.q_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6aacfee4be69b35559fdee27d2563412433f4bd6af70ef9fca56a428a58dc8d
3
+ size 8766368
out_tensor/model.layers.0.self_attn.v_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b48d2280c745ef92eaaf5640fad72c101613b88aa0effe0b67b239583b76923c
3
+ size 8881440
out_tensor/model.layers.1.mlp.down_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c36a73be58b884d6b0d406b7aaa9061066a189e41522232cc70e0bb9669ab97
3
+ size 25048616
out_tensor/model.layers.1.mlp.gate_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8e59b26e8bff000d92534b0168d779a45c953ebf740754c1e8c5c4f5a57f0e6
3
+ size 23839000
out_tensor/model.layers.1.mlp.up_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86d591b8804a110148bea485d33a4f14fd00dd5616a064e9487f7fd3b6eda22b
3
+ size 24675600
out_tensor/model.layers.1.self_attn.k_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:644f2496a3bbf754352920cf4063bd1622110e953455ab2b694492b476e61852
3
+ size 10978600
out_tensor/model.layers.1.self_attn.o_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79001786ac13e199ffa842803e99318471d109b83f92534eb256405c93798120
3
+ size 10978600
out_tensor/model.layers.1.self_attn.q_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b44ebdcf5968f4c8334fe04af1da3c89443d08fdd8ae4f3a02987b8a2be32fc5
3
+ size 10978600
out_tensor/model.layers.1.self_attn.v_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b7c1da24826c5dc0ac466dd4cfcdcbb5e37d0348900068b13f17b16030bd337
3
+ size 12862760
out_tensor/model.layers.10.mlp.down_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9708845897428da99c5fef71c7c6c26df2f45b18d3e5ebc675fbae7c05b2c389
3
+ size 25048624
out_tensor/model.layers.10.mlp.gate_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4de879a37787ee49721c8d711acc333dc5af83ce05ce1845340384004f4b09d
3
+ size 23839008
out_tensor/model.layers.10.mlp.up_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0598ac643c5984df4f5cd2012f6682b5e5a4266b40d7ba3b4e70b69b0338723
3
+ size 24675600
out_tensor/model.layers.10.self_attn.k_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5306ec8dab319218fb9663b4dfc3d46cd232d256bdfc66f10a97adf490c19bc6
3
+ size 4473760
out_tensor/model.layers.10.self_attn.o_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f235b7633698a0f273ffa6292d1a3dc3ecb952b51cc85435139aded45e4b6184
3
+ size 4473760
out_tensor/model.layers.10.self_attn.q_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d52f703a09df4833c8ed0d46ee0b0734e5658acc6632854eecf03a55e30bc385
3
+ size 4473760
out_tensor/model.layers.10.self_attn.v_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:875ee32a475329c7c8ee6b6f061e9b4aeb3fc7b8797b8f479e2691b77d0962c5
3
+ size 4473760
out_tensor/model.layers.11.mlp.down_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:816fb8f05ebe65dc252c5db573746ed97560d7b27e4e1ed00b3478502e063900
3
+ size 23974008
out_tensor/model.layers.11.mlp.gate_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ee36f3604157d6319cc2c1155d37197d4c0372933b1785c8b703a6d1b624499
3
+ size 22737624
out_tensor/model.layers.11.mlp.up_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17f2d21eaebab90f7e87fa6854b813232c7aa650eb1a0ac70c8f3884966950e6
3
+ size 23266576
out_tensor/model.layers.11.self_attn.k_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b86d4f247f57932c3ebfe7a6e5b7a15698c8fa07b2b0bc4f829f22b868352082
3
+ size 4473760
out_tensor/model.layers.11.self_attn.o_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:560fa25340c3ace64f3253dfb9b37c3794e7cdc382e404501e0b2a309cf0be52
3
+ size 4473760
out_tensor/model.layers.11.self_attn.q_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc589f35a5a0d4b27bec00d59411d083eb44d1ca29f6b63a4f718ad5e1115c91
3
+ size 4473760
out_tensor/model.layers.11.self_attn.v_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a3c1faba688ddeb2cadad78527fe4a40622b644edaaffb5c0476994ecd2c733
3
+ size 4473760
out_tensor/model.layers.12.mlp.down_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c418a01203a0de4484f6bbcd1f99a2bf70fee86f77d104d0fd32ced80065e95
3
+ size 25048624
out_tensor/model.layers.12.mlp.gate_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6a3459d8f824d99d1778702bf3047cc8a575b3bab0654abf568c3fd05195f6f
3
+ size 23839008
out_tensor/model.layers.12.mlp.up_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb113bce78227454218f9914588e80e2cbc1848be8a14324b9d0115208203277
3
+ size 24675600
out_tensor/model.layers.12.self_attn.k_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d951d64cb90e44dee8dfb7f8d1c653a27dd7403201dac344517a674e9647981e
3
+ size 4572064
out_tensor/model.layers.12.self_attn.o_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f62db4a948c5fd7760961fae13d383c08092168bc1d2f6cbf3732b5c66361b9e
3
+ size 4572064
out_tensor/model.layers.12.self_attn.q_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc060607b1ce6f66bed8c62019675e2ffb2b2b3adc430f9d05fe4bcdd50fdc74
3
+ size 4572064
out_tensor/model.layers.12.self_attn.v_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6f93b0d13c2aced2c2d46913b55f1bf99e6aa81c4119fb64d0c4186900be36d
3
+ size 4866976
out_tensor/model.layers.13.mlp.down_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:447d6fa976edbbca31377be95fa8a65079bfeaea11b1868e549777edacf20f0b
3
+ size 24475184
out_tensor/model.layers.13.mlp.gate_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f6849a263686ea0234eb0f10b308ff5d1e55aa96f8b80469b91c23447333a20
3
+ size 23266592
out_tensor/model.layers.13.mlp.up_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01e8c831384957f8b7e54dc8ed7f8a1432bbf2b5dd97a67a291f440615a5e4ca
3
+ size 23266576
out_tensor/model.layers.13.self_attn.k_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daea163599f614595640da330768bde7deaf2a45387367319490ff4fab5af2ba
3
+ size 4572064
out_tensor/model.layers.13.self_attn.o_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f04238d500245e59e8b375c97fbef3cc7b52d9774c7d764d60b80627cb010362
3
+ size 4572064
out_tensor/model.layers.13.self_attn.q_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf46fd0b1dbdc5d7fa1fcb89214475faf6ca1f543f821b58d8a7341d166b8a50
3
+ size 4572064
out_tensor/model.layers.13.self_attn.v_proj.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bed739b8f640b7f51c31ae778f3229678e7a49b84bc14f4f2b5d041aed52015b
3
+ size 6636256