Zaid commited on
Commit
5112867
1 Parent(s): 2fb81a4

add diacritizer

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. poetry_diacritizer/__init__.py +1 -0
  2. poetry_diacritizer/__pycache__/__init__.cpython-310.pyc +0 -0
  3. poetry_diacritizer/__pycache__/config_manager.cpython-310.pyc +0 -0
  4. poetry_diacritizer/__pycache__/config_manager.cpython-38.pyc +0 -0
  5. poetry_diacritizer/__pycache__/dataset.cpython-310.pyc +0 -0
  6. poetry_diacritizer/__pycache__/dataset.cpython-38.pyc +0 -0
  7. poetry_diacritizer/__pycache__/diacritizer.cpython-310.pyc +0 -0
  8. poetry_diacritizer/__pycache__/diacritizer.cpython-38.pyc +0 -0
  9. poetry_diacritizer/__pycache__/gpt_model.cpython-38.pyc +0 -0
  10. poetry_diacritizer/__pycache__/options.cpython-310.pyc +0 -0
  11. poetry_diacritizer/__pycache__/options.cpython-38.pyc +0 -0
  12. poetry_diacritizer/__pycache__/predict.cpython-310.pyc +0 -0
  13. poetry_diacritizer/__pycache__/tester.cpython-38.pyc +0 -0
  14. poetry_diacritizer/__pycache__/trainer.cpython-38.pyc +0 -0
  15. poetry_diacritizer/config/ashaar.yml +52 -0
  16. poetry_diacritizer/config/baseline.yml +47 -0
  17. poetry_diacritizer/config/cbhg.yml +52 -0
  18. poetry_diacritizer/config/cbhg2.yml +51 -0
  19. poetry_diacritizer/config/gpt-0.yml +46 -0
  20. poetry_diacritizer/config/gpt-1.yml +46 -0
  21. poetry_diacritizer/config/gpt-2.yml +46 -0
  22. poetry_diacritizer/config/gpt-3.yml +46 -0
  23. poetry_diacritizer/config/gpt-4.yml +46 -0
  24. poetry_diacritizer/config/gpt-5.yml +46 -0
  25. poetry_diacritizer/config/gpt-6.yml +46 -0
  26. poetry_diacritizer/config/gpt-7.yml +46 -0
  27. poetry_diacritizer/config/gpt-8.yml +46 -0
  28. poetry_diacritizer/config/gpt-9.yml +46 -0
  29. poetry_diacritizer/config/gpt-cls-0-tash-proc.yml +46 -0
  30. poetry_diacritizer/config/gpt-cls-0-test.yml +46 -0
  31. poetry_diacritizer/config/gpt-cls-0.yml +46 -0
  32. poetry_diacritizer/config/gpt-cls-1-tash-proc.yml +46 -0
  33. poetry_diacritizer/config/gpt-cls-1.yml +46 -0
  34. poetry_diacritizer/config/gpt-cls-2-tash-proc.yml +46 -0
  35. poetry_diacritizer/config/gpt-cls-2.yml +46 -0
  36. poetry_diacritizer/config/gpt-cls-3-tash-proc.yml +46 -0
  37. poetry_diacritizer/config/gpt-cls-3.yml +46 -0
  38. poetry_diacritizer/config/gpt-cls-4-tash-proc.yml +46 -0
  39. poetry_diacritizer/config/gpt-cls-4.yml +46 -0
  40. poetry_diacritizer/config/gpt-cls-5-tash-proc.yml +46 -0
  41. poetry_diacritizer/config/gpt-cls-5-test.yml +46 -0
  42. poetry_diacritizer/config/gpt-cls-5.yml +46 -0
  43. poetry_diacritizer/config/gpt-cls-6-tash-proc.yml +46 -0
  44. poetry_diacritizer/config/gpt-cls-6.yml +46 -0
  45. poetry_diacritizer/config/gpt-cls-7-tash-proc.yml +46 -0
  46. poetry_diacritizer/config/gpt-cls-7.yml +46 -0
  47. poetry_diacritizer/config/gpt-cls-8-tash-proc.yml +46 -0
  48. poetry_diacritizer/config/gpt-cls-8.yml +46 -0
  49. poetry_diacritizer/config/gpt-cls-9-tash-proc.yml +46 -0
  50. poetry_diacritizer/config/gpt-cls-9-test.yml +46 -0
poetry_diacritizer/__init__.py ADDED
@@ -0,0 +1 @@
 
 
1
+ from poetry_diacritizer import predict
poetry_diacritizer/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (210 Bytes). View file
 
poetry_diacritizer/__pycache__/config_manager.cpython-310.pyc ADDED
Binary file (10.7 kB). View file
 
poetry_diacritizer/__pycache__/config_manager.cpython-38.pyc ADDED
Binary file (10.4 kB). View file
 
poetry_diacritizer/__pycache__/dataset.cpython-310.pyc ADDED
Binary file (6.5 kB). View file
 
poetry_diacritizer/__pycache__/dataset.cpython-38.pyc ADDED
Binary file (6.5 kB). View file
 
poetry_diacritizer/__pycache__/diacritizer.cpython-310.pyc ADDED
Binary file (2.87 kB). View file
 
poetry_diacritizer/__pycache__/diacritizer.cpython-38.pyc ADDED
Binary file (3.27 kB). View file
 
poetry_diacritizer/__pycache__/gpt_model.cpython-38.pyc ADDED
Binary file (8.17 kB). View file
 
poetry_diacritizer/__pycache__/options.cpython-310.pyc ADDED
Binary file (1.34 kB). View file
 
poetry_diacritizer/__pycache__/options.cpython-38.pyc ADDED
Binary file (1.4 kB). View file
 
poetry_diacritizer/__pycache__/predict.cpython-310.pyc ADDED
Binary file (5.78 kB). View file
 
poetry_diacritizer/__pycache__/tester.cpython-38.pyc ADDED
Binary file (2.12 kB). View file
 
poetry_diacritizer/__pycache__/trainer.cpython-38.pyc ADDED
Binary file (12 kB). View file
 
poetry_diacritizer/config/ashaar.yml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ session_name: base
2
+
3
+ data_directory: "data"
4
+ data_type: "ashaar_proc"
5
+ log_directory: "log_dir_ashaar"
6
+ load_training_data: true
7
+ load_test_data: false
8
+ load_validation_data: true
9
+ n_training_examples: null # null load all training examples, good for fast loading
10
+ n_test_examples: null # null load all test examples
11
+ n_validation_examples: null # null load all validation examples
12
+ test_file_name: "test.csv"
13
+ is_data_preprocessed: false # The data file is organized as (original text | text | diacritics)
14
+ data_separator: '|' # Required if the data already processed
15
+ diacritics_separator: '*' # Required if the data already processed
16
+ text_encoder: ArabicEncoderWithStartSymbol
17
+ text_cleaner: valid_arabic_cleaners # a white list that uses only Arabic letters, punctuations, and a space
18
+ max_len: 600 # sentences larger than this size will not be used
19
+ max_sen_len: null
20
+
21
+ max_steps: 10000
22
+ learning_rate: 0.001
23
+ batch_size: 32
24
+ adam_beta1: 0.9
25
+ adam_beta2: 0.999
26
+ use_decay: true
27
+ weight_decay: 0.0
28
+ embedding_dim: 256
29
+ use_prenet: false
30
+ prenet_sizes: [512, 256]
31
+ cbhg_projections: [128, 256]
32
+ cbhg_filters: 16
33
+ cbhg_gru_units: 256
34
+ post_cbhg_layers_units: [256, 256]
35
+ post_cbhg_use_batch_norm: true
36
+
37
+ use_mixed_precision: false
38
+ optimizer_type: Adam
39
+ device: cuda
40
+
41
+ # LOGGING
42
+ evaluate_frequency: 50000000
43
+ max_eval_batches: 100
44
+ evaluate_with_error_rates_frequency: 1000
45
+ n_predicted_text_tensorboard: 10 # To be written to the tensorboard
46
+ model_save_frequency: 1000
47
+ train_plotting_frequency: 50000000 # No plotting for this model
48
+ n_steps_avg_losses: [100, 500, 1_000, 5_000] # command line display of average loss values for the last n steps
49
+ error_rates_n_batches: 10000 # if calculating error rate is slow, then you can specify the number of batches to be calculated
50
+
51
+ test_model_path: null # load the last saved model
52
+ train_resume_model_path: null # load last saved model
poetry_diacritizer/config/baseline.yml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ session_name: base
2
+
3
+ data_directory: "data"
4
+ data_type: "CA_MSA"
5
+ log_directory: "log_dir"
6
+ load_training_data: true
7
+ load_test_data: false
8
+ load_validation_data: true
9
+ n_training_examples: null # null load all training examples, good for fast loading
10
+ n_test_examples: null # null load all test examples
11
+ n_validation_examples: null # null load all validation examples
12
+ test_file_name: "test.csv"
13
+ is_data_preprocessed: false # The data file is organized as (original text | text | diacritics)
14
+ data_separator: '|' # Required if the data already processed
15
+ diacritics_separator: '*' # Required if the data already processed
16
+ text_encoder: ArabicEncoderWithStartSymbol
17
+ text_cleaner: valid_arabic_cleaners # a white list that uses only Arabic letters, punctuations, and a space
18
+ max_len: 600 # sentences larger than this size will not be used
19
+
20
+
21
+ max_steps: 2_000_000
22
+ learning_rate: 0.001
23
+ batch_size: 64
24
+ adam_beta1: 0.9
25
+ adam_beta2: 0.999
26
+ use_decay: true
27
+ weight_decay: 0.0
28
+ embedding_dim: 512
29
+ n_layers: 3
30
+ layers_units: [256, 256, 256]
31
+ use_mixed_precision: false
32
+ optimizer_type: Adam
33
+ use_batch_norm: False
34
+ device: cuda
35
+ max_sen_len: 256
36
+
37
+ # LOGGING
38
+ evaluate_frequency: 5000
39
+ evaluate_with_error_rates_frequency: 5000
40
+ n_predicted_text_tensorboard: 10 # To be written to the tensorboard
41
+ model_save_frequency: 5000
42
+ train_plotting_frequency: 50000000 # No plotting for this model
43
+ n_steps_avg_losses: [100, 500, 1_000, 5_000] # command line display of average loss values for the last n steps
44
+ error_rates_n_batches: 10000 # if calculating error rate is slow, then you can specify the number of batches to be calculated
45
+
46
+ test_model_path: null # load the last saved model
47
+ train_resume_model_path: null # load last saved model
poetry_diacritizer/config/cbhg.yml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ session_name: base
2
+
3
+ data_directory: "data"
4
+ data_type: "CA_MSA"
5
+ log_directory: "log_dir_cbhg"
6
+ load_training_data: true
7
+ load_test_data: false
8
+ load_validation_data: true
9
+ n_training_examples: null # null load all training examples, good for fast loading
10
+ n_test_examples: null # null load all test examples
11
+ n_validation_examples: null # null load all validation examples
12
+ test_file_name: "test.csv"
13
+ is_data_preprocessed: false # The data file is organized as (original text | text | diacritics)
14
+ data_separator: '|' # Required if the data already processed
15
+ diacritics_separator: '*' # Required if the data already processed
16
+ text_encoder: ArabicEncoderWithStartSymbol
17
+ text_cleaner: valid_arabic_cleaners # a white list that uses only Arabic letters, punctuations, and a space
18
+ max_len: 600 # sentences larger than this size will not be used
19
+ max_sen_len: null
20
+
21
+ max_steps: 5000
22
+ learning_rate: 0.001
23
+ batch_size: 32
24
+ adam_beta1: 0.9
25
+ adam_beta2: 0.999
26
+ use_decay: true
27
+ weight_decay: 0.0
28
+ embedding_dim: 256
29
+ use_prenet: false
30
+ prenet_sizes: [512, 256]
31
+ cbhg_projections: [128, 256]
32
+ cbhg_filters: 16
33
+ cbhg_gru_units: 256
34
+ post_cbhg_layers_units: [256, 256]
35
+ post_cbhg_use_batch_norm: true
36
+
37
+ use_mixed_precision: false
38
+ optimizer_type: Adam
39
+ device: cuda
40
+
41
+ # LOGGING
42
+ evaluate_frequency: 50000000
43
+ max_eval_batches: 100
44
+ evaluate_with_error_rates_frequency: 1000
45
+ n_predicted_text_tensorboard: 10 # To be written to the tensorboard
46
+ model_save_frequency: 5000
47
+ train_plotting_frequency: 50000000 # No plotting for this model
48
+ n_steps_avg_losses: [100, 500, 1_000, 5_000] # command line display of average loss values for the last n steps
49
+ error_rates_n_batches: 10000 # if calculating error rate is slow, then you can specify the number of batches to be calculated
50
+
51
+ test_model_path: null # load the last saved model
52
+ train_resume_model_path: null # load last saved model
poetry_diacritizer/config/cbhg2.yml ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ session_name: base
2
+
3
+ data_directory: "ashaar"
4
+ data_type: "CA_MSA"
5
+ log_directory: "/content/drive/MyDrive/Research/Barmajan/Diacritization/log_ashaar_dir"
6
+ load_training_data: true
7
+ load_test_data: false
8
+ load_validation_data: true
9
+ n_training_examples: null # null load all training examples, good for fast loading
10
+ n_test_examples: null # null load all test examples
11
+ n_validation_examples: null # null load all validation examples
12
+ test_file_name: "test.csv"
13
+ is_data_preprocessed: false # The data file is organized as (original text | text | diacritics)
14
+ data_separator: '|' # Required if the data already processed
15
+ diacritics_separator: '*' # Required if the data already processed
16
+ text_encoder: ArabicEncoderWithStartSymbol
17
+ text_cleaner: valid_arabic_cleaners # a white list that uses only Arabic letters, punctuations, and a space
18
+ max_len: 600 # sentences larger than this size will not be used
19
+
20
+
21
+ max_steps: 25_000
22
+ learning_rate: 0.001
23
+ batch_size: 32
24
+ adam_beta1: 0.9
25
+ adam_beta2: 0.999
26
+ use_decay: true
27
+ weight_decay: 0.0
28
+ embedding_dim: 256
29
+ use_prenet: false
30
+ prenet_sizes: [512, 256]
31
+ cbhg_projections: [128, 256]
32
+ cbhg_filters: 16
33
+ cbhg_gru_units: 256
34
+ post_cbhg_layers_units: [256, 256]
35
+ post_cbhg_use_batch_norm: true
36
+
37
+ use_mixed_precision: false
38
+ optimizer_type: Adam
39
+ device: cuda
40
+
41
+ # LOGGING
42
+ evaluate_frequency: 1000
43
+ evaluate_with_error_rates_frequency: 1000
44
+ n_predicted_text_tensorboard: 10 # To be written to the tensorboard
45
+ model_save_frequency: 1000
46
+ train_plotting_frequency: 50000000 # No plotting for this model
47
+ n_steps_avg_losses: [100, 500, 1_000, 5_000] # command line display of average loss values for the last n steps
48
+ error_rates_n_batches: 10000 # if calculating error rate is slow, then you can specify the number of batches to be calculated
49
+
50
+ test_model_path: null # load the last saved model
51
+ train_resume_model_path: "/content/drive/MyDrive/Research/Barmajan/Diacritization/log_cleaned_dir/CA_MSA.base.cbhg/models/20000-snapshot.pt" # load last saved model
poetry_diacritizer/config/gpt-0.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_0
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 0
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: true
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-1.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_1
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 1
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: true
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-2.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_2
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 2
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: true
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-3.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_3
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 3
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: true
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-4.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_4
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 4
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: true
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-5.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_5
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 5
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: true
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-6.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_6
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 6
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: true
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-7.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_7
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 7
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: true
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-8.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_8
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 8
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: true
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-9.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_9
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 9
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: true
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-0-tash-proc.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: tash_proc
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_0_tash_proc
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 0
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-0-test.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_0_test
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 0
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-0.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_0
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 0
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-1-tash-proc.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: tash_proc
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_1_tash_proc
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 1
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-1.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_1
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 1
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-2-tash-proc.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: tash_proc
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_2_tash_proc
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 2
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-2.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_2
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 2
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-3-tash-proc.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: tash_proc
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_3_tash_proc
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 3
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-3.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_3
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 3
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-4-tash-proc.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: tash_proc
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_4_tash_proc
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 4
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-4.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_4
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 4
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-5-tash-proc.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: tash_proc
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_5_tash_proc
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 5
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-5-test.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: logs/log_dir_cls_5_test
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 5
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-5.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_5
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 5
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-6-tash-proc.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: tash_proc
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_6_tash_proc
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 6
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-6.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_6
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 6
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-7-tash-proc.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: tash_proc
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_7_tash_proc
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 7
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-7.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_7
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 7
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-8-tash-proc.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: tash_proc
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_8_tash_proc
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 8
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-8.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_8
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 8
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-9-tash-proc.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: tash_proc
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: log_dir_cls_9_tash_proc
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 9
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0
poetry_diacritizer/config/gpt-cls-9-test.yml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ base_model_path: ashaar-from-scratch-with-spaces-no-tatweel-epochs-75
4
+ batch_size: 64
5
+ data_directory: data
6
+ data_separator: '|'
7
+ data_type: CA_MSA
8
+ device: cuda
9
+ diacritics_separator: '*'
10
+ error_rates_n_batches: 10000
11
+ evaluate_frequency: 50000000
12
+ evaluate_with_error_rates_frequency: 1000
13
+ freeze: true
14
+ is_data_preprocessed: false
15
+ learning_rate: 0.001
16
+ load_test_data: false
17
+ load_training_data: true
18
+ load_validation_data: true
19
+ log_directory: logs/log_dir_cls_9_test
20
+ max_eval_batches: -1
21
+ max_len: 600
22
+ max_sen_len: 256
23
+ max_steps: 5000
24
+ model_save_frequency: 5000
25
+ n_layer: 9
26
+ n_predicted_text_tensorboard: 10
27
+ n_steps_avg_losses:
28
+ - 100
29
+ - 500
30
+ - 1000
31
+ - 5000
32
+ n_test_examples: null
33
+ n_training_examples: null
34
+ n_validation_examples: null
35
+ optimizer_type: Adam
36
+ session_name: base
37
+ test_file_name: test.csv
38
+ test_model_path: null
39
+ text_cleaner: valid_arabic_cleaners
40
+ text_encoder: ArabicEncoderWithStartSymbol
41
+ train_plotting_frequency: 50000000
42
+ train_resume_model_path: null
43
+ use_decay: true
44
+ use_lstm: false
45
+ use_mixed_precision: false
46
+ weight_decay: 0.0