abuelnasr commited on
Commit
4bef79c
1 Parent(s): 5223075

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: keras-nlp
3
+ ---
4
+ This is a [`Phi3` model](https://keras.io/api/keras_nlp/models/phi3) uploaded using the KerasNLP library and can be used with JAX, TensorFlow, and PyTorch backends.
5
+ Model config:
6
+ * **name:** phi3_backbone_1
7
+ * **trainable:** True
8
+ * **vocabulary_size:** 32064
9
+ * **num_layers:** 32
10
+ * **num_query_heads:** 32
11
+ * **hidden_dim:** 3072
12
+ * **intermediate_dim:** 8192
13
+ * **num_key_value_heads:** 32
14
+ * **layer_norm_epsilon:** 1e-05
15
+ * **dropout:** 0.0
16
+ * **max_sequence_length:** 131072
17
+ * **original_max_sequence_length:** 4096
18
+ * **rope_max_wavelength:** 10000.0
19
+ * **rope_scaling_type:** su
20
+ * **rope_scaling_short_factor:** [1.05, 1.05, 1.05, 1.1, 1.1, 1.1500000000000001, 1.2000000000000002, 1.2500000000000002, 1.3000000000000003, 1.3500000000000003, 1.5000000000000004, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.000000000000001, 2.0500000000000007, 2.0500000000000007, 2.0500000000000007, 2.1000000000000005, 2.1000000000000005, 2.1000000000000005, 2.1500000000000004, 2.1500000000000004, 2.3499999999999996, 2.549999999999999, 2.5999999999999988, 2.5999999999999988, 2.7499999999999982, 2.849999999999998, 2.849999999999998, 2.9499999999999975]
21
+ * **rope_scaling_long_factor:** [1.0299999713897705, 1.0499999523162842, 1.0499999523162842, 1.0799999237060547, 1.2299998998641968, 1.2299998998641968, 1.2999999523162842, 1.4499999284744263, 1.5999999046325684, 1.6499998569488525, 1.8999998569488525, 2.859999895095825, 3.68999981880188, 5.419999599456787, 5.489999771118164, 5.489999771118164, 9.09000015258789, 11.579999923706055, 15.65999984741211, 15.769999504089355, 15.789999961853027, 18.360000610351562, 21.989999771118164, 23.079999923706055, 30.009998321533203, 32.35000228881836, 32.590003967285156, 35.56000518798828, 39.95000457763672, 53.840003967285156, 56.20000457763672, 57.95000457763672, 59.29000473022461, 59.77000427246094, 59.920005798339844, 61.190006256103516, 61.96000671386719, 62.50000762939453, 63.3700065612793, 63.48000717163086, 63.48000717163086, 63.66000747680664, 63.850006103515625, 64.08000946044922, 64.760009765625, 64.80001068115234, 64.81001281738281, 64.81001281738281]
22
+
23
+ This model card has been generated automatically and should be completed by the model author. See [Model Cards documentation](https://huggingface.co/docs/hub/model-cards) for more information.
assets/tokenizer/vocabulary.spm ADDED
Binary file (500 kB). View file
 
config.json ADDED
@@ -0,0 +1,121 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "module": "keras_nlp.src.models.phi3.phi3_backbone",
3
+ "class_name": "Phi3Backbone",
4
+ "config": {
5
+ "name": "phi3_backbone_1",
6
+ "trainable": true,
7
+ "vocabulary_size": 32064,
8
+ "num_layers": 32,
9
+ "num_query_heads": 32,
10
+ "hidden_dim": 3072,
11
+ "intermediate_dim": 8192,
12
+ "num_key_value_heads": 32,
13
+ "layer_norm_epsilon": 1e-05,
14
+ "dropout": 0.0,
15
+ "max_sequence_length": 131072,
16
+ "original_max_sequence_length": 4096,
17
+ "rope_max_wavelength": 10000.0,
18
+ "rope_scaling_type": "su",
19
+ "rope_scaling_short_factor": [
20
+ 1.05,
21
+ 1.05,
22
+ 1.05,
23
+ 1.1,
24
+ 1.1,
25
+ 1.1500000000000001,
26
+ 1.2000000000000002,
27
+ 1.2500000000000002,
28
+ 1.3000000000000003,
29
+ 1.3500000000000003,
30
+ 1.5000000000000004,
31
+ 2.000000000000001,
32
+ 2.000000000000001,
33
+ 2.000000000000001,
34
+ 2.000000000000001,
35
+ 2.000000000000001,
36
+ 2.000000000000001,
37
+ 2.000000000000001,
38
+ 2.000000000000001,
39
+ 2.000000000000001,
40
+ 2.000000000000001,
41
+ 2.000000000000001,
42
+ 2.000000000000001,
43
+ 2.000000000000001,
44
+ 2.000000000000001,
45
+ 2.000000000000001,
46
+ 2.000000000000001,
47
+ 2.000000000000001,
48
+ 2.000000000000001,
49
+ 2.000000000000001,
50
+ 2.000000000000001,
51
+ 2.000000000000001,
52
+ 2.0500000000000007,
53
+ 2.0500000000000007,
54
+ 2.0500000000000007,
55
+ 2.1000000000000005,
56
+ 2.1000000000000005,
57
+ 2.1000000000000005,
58
+ 2.1500000000000004,
59
+ 2.1500000000000004,
60
+ 2.3499999999999996,
61
+ 2.549999999999999,
62
+ 2.5999999999999988,
63
+ 2.5999999999999988,
64
+ 2.7499999999999982,
65
+ 2.849999999999998,
66
+ 2.849999999999998,
67
+ 2.9499999999999975
68
+ ],
69
+ "rope_scaling_long_factor": [
70
+ 1.0299999713897705,
71
+ 1.0499999523162842,
72
+ 1.0499999523162842,
73
+ 1.0799999237060547,
74
+ 1.2299998998641968,
75
+ 1.2299998998641968,
76
+ 1.2999999523162842,
77
+ 1.4499999284744263,
78
+ 1.5999999046325684,
79
+ 1.6499998569488525,
80
+ 1.8999998569488525,
81
+ 2.859999895095825,
82
+ 3.68999981880188,
83
+ 5.419999599456787,
84
+ 5.489999771118164,
85
+ 5.489999771118164,
86
+ 9.09000015258789,
87
+ 11.579999923706055,
88
+ 15.65999984741211,
89
+ 15.769999504089355,
90
+ 15.789999961853027,
91
+ 18.360000610351562,
92
+ 21.989999771118164,
93
+ 23.079999923706055,
94
+ 30.009998321533203,
95
+ 32.35000228881836,
96
+ 32.590003967285156,
97
+ 35.56000518798828,
98
+ 39.95000457763672,
99
+ 53.840003967285156,
100
+ 56.20000457763672,
101
+ 57.95000457763672,
102
+ 59.29000473022461,
103
+ 59.77000427246094,
104
+ 59.920005798339844,
105
+ 61.190006256103516,
106
+ 61.96000671386719,
107
+ 62.50000762939453,
108
+ 63.3700065612793,
109
+ 63.48000717163086,
110
+ 63.48000717163086,
111
+ 63.66000747680664,
112
+ 63.850006103515625,
113
+ 64.08000946044922,
114
+ 64.760009765625,
115
+ 64.80001068115234,
116
+ 64.81001281738281,
117
+ 64.81001281738281
118
+ ]
119
+ },
120
+ "registered_name": "keras_nlp>Phi3Backbone"
121
+ }
metadata.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "keras_version": "3.3.3",
3
+ "keras_nlp_version": "0.10.0",
4
+ "parameter_count": 3821079552,
5
+ "date_saved": "2024-05-06@23:42:02"
6
+ }
model.weights.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7c2d6ff5183500e9167a354765ed755ab633417e989acf9e8c4aa07a2ea9e60
3
+ size 7643265592
tokenizer.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "module": "keras_nlp.src.models.phi3.phi3_tokenizer",
3
+ "class_name": "Phi3Tokenizer",
4
+ "config": {
5
+ "name": "phi3_tokenizer",
6
+ "trainable": true,
7
+ "dtype": "int32",
8
+ "proto": null,
9
+ "sequence_length": null
10
+ },
11
+ "registered_name": "keras_nlp>Phi3Tokenizer"
12
+ }