Romain1311 commited on
Commit
a90c94d
1 Parent(s): 493d30c

Training in progress, step 500

Browse files
Files changed (3) hide show
  1. config.json +14 -16
  2. model.safetensors +2 -2
  3. tokenizer.json +1 -6
config.json CHANGED
@@ -1,14 +1,13 @@
1
  {
2
- "_name_or_path": "transformersbook/bert-base-uncased-finetuned-clinc",
 
3
  "architectures": [
4
- "BertForSequenceClassification"
5
  ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
- "gradient_checkpointing": false,
9
- "hidden_act": "gelu",
10
- "hidden_dropout_prob": 0.1,
11
- "hidden_size": 768,
12
  "id2label": {
13
  "0": "restaurant_reviews",
14
  "1": "nutrition_info",
@@ -163,7 +162,6 @@
163
  "150": "change_volume"
164
  },
165
  "initializer_range": 0.02,
166
- "intermediate_size": 3072,
167
  "label2id": {
168
  "accept_reservations": 9,
169
  "account_blocked": 2,
@@ -317,17 +315,17 @@
317
  "who_made_you": 96,
318
  "yes": 107
319
  },
320
- "layer_norm_eps": 1e-12,
321
  "max_position_embeddings": 512,
322
- "model_type": "bert",
323
- "num_attention_heads": 12,
324
- "num_hidden_layers": 12,
325
  "pad_token_id": 0,
326
- "position_embedding_type": "absolute",
327
  "problem_type": "single_label_classification",
 
 
 
 
328
  "torch_dtype": "float32",
329
  "transformers_version": "4.39.3",
330
- "type_vocab_size": 2,
331
- "use_cache": true,
332
  "vocab_size": 30522
333
  }
 
1
  {
2
+ "_name_or_path": "distilbert-base-uncased",
3
+ "activation": "gelu",
4
  "architectures": [
5
+ "DistilBertForSequenceClassification"
6
  ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
 
 
11
  "id2label": {
12
  "0": "restaurant_reviews",
13
  "1": "nutrition_info",
 
162
  "150": "change_volume"
163
  },
164
  "initializer_range": 0.02,
 
165
  "label2id": {
166
  "accept_reservations": 9,
167
  "account_blocked": 2,
 
315
  "who_made_you": 96,
316
  "yes": 107
317
  },
 
318
  "max_position_embeddings": 512,
319
+ "model_type": "distilbert",
320
+ "n_heads": 12,
321
+ "n_layers": 6,
322
  "pad_token_id": 0,
 
323
  "problem_type": "single_label_classification",
324
+ "qa_dropout": 0.1,
325
+ "seq_classif_dropout": 0.2,
326
+ "sinusoidal_pos_embds": false,
327
+ "tie_weights_": true,
328
  "torch_dtype": "float32",
329
  "transformers_version": "4.39.3",
 
 
330
  "vocab_size": 30522
331
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ae0cde19a3ccf1842be67bb2f9e5ac0d15db8d0eb22df4fad2a53de3792e349
3
- size 438416972
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:577de16c52f88b45cbda69ff39ce5d835270d4474331fd2e0568e3c60c29443a
3
+ size 268290900
tokenizer.json CHANGED
@@ -1,11 +1,6 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 512,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
  "padding": null,
10
  "added_tokens": [
11
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {