ashabrawy commited on
Commit
b779bb6
1 Parent(s): d778192

jais-family-590m_gamma_0.0_0_lamda_0.0-0_English_sft/

Browse files
README.md CHANGED
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/afz225/uncategorized/runs/x7hgncrn)
31
 
32
  This model was trained with SFT.
33
 
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/afz225/uncategorized/runs/b7i0h42r)
31
 
32
  This model was trained with SFT.
33
 
adapter_config.json CHANGED
@@ -23,14 +23,14 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "down_proj",
27
  "k_proj",
28
- "q_proj",
29
  "v_proj",
 
 
30
  "up_proj",
31
- "lm_head",
32
- "gate_proj",
33
- "o_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "lm_head",
27
  "k_proj",
28
+ "o_proj",
29
  "v_proj",
30
+ "q_proj",
31
+ "down_proj",
32
  "up_proj",
33
+ "gate_proj"
 
 
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "use_dora": false,
global_step110/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:148341ada6d6c067d79601e1320b44f907226ac3efeae77fb8fcfd77b602dcea
3
+ size 160351592
global_step110/zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59b46b298d38abf3665ab5161eedabbc52547342d1a243df07e089603f006048
3
+ size 2080173
global_step110/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f568895fefa0bd71f02f02e2add2c0c0b32b33804e1b2e25ad15125beeab43b0
3
+ size 160351464
global_step110/zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5886183566f3658a5a0277a4d809bf5655d2fc32a664617e677a9f08da6590e7
3
+ size 2080173
global_step110/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a20a4d9edd99013dd89681dc93ba801cc1e2ab01cc6c6fd5bfd8b5656fa7d849
3
+ size 160351464
global_step110/zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ddd8348c3375c7faf885e530a43d82ea4c83f577b870468c3172065540db9ab
3
+ size 2080173
global_step110/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ea61af521dd0685aa444967c1093efabd818e0a305b9785cb1bb5a50a50cdb2
3
+ size 160351464
global_step110/zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87fd31b11740c891db60bb65e28ab86aa827a70ef7f1c62aeabbd1749f9fefd9
3
+ size 2080173
global_step110/zero_pp_rank_4_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0e3ca859e910b924306b12ef56b4a6897fee18e9c8ab6dc2db5c2d7487a08a6
3
+ size 160351464
global_step110/zero_pp_rank_4_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7feaf65961b7028d1204e1f516a6e735a42b7c121a2a19b75b35a8fc99a5242c
3
+ size 2080173
global_step110/zero_pp_rank_5_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d46014c2dd75a625e265286f78607195f70a15609a0194c726e2d3e2198f726b
3
+ size 160351464
global_step110/zero_pp_rank_5_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1944b6dfa554757195572a6a8a58a86d7aeceea5909b414841bdf2e7fde1d6a6
3
+ size 2080173
global_step110/zero_pp_rank_6_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc5d3491397d8e5a8ce46a0885dfe3338b95517803238443e2b4498fd1f15da2
3
+ size 160351464
global_step110/zero_pp_rank_6_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45039267c53264ada76cd967ead5f2f99a2d5582df6ff90ab030534e9789275c
3
+ size 2080173
global_step110/zero_pp_rank_7_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:524850da392f76e055b91473d95806b529c4b2511bbf32d7d3f40be68c54da33
3
+ size 160351464
global_step110/zero_pp_rank_7_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af4dcd3b941d7374339665937174775e29c4761236af92c69b7a9eeda5997a1e
3
+ size 2080173
latest CHANGED
@@ -1 +1 @@
1
- global_step220
 
1
+ global_step110
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61360da61b8698b89e39fcbad0f2ff60303e873ac32dd21b7791fd70e7103c54
3
  size 6968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46dcc2c6ea99d2ed1282841aec4f717daa86f477e1b5879581d6e68cccef4624
3
  size 6968