aboutsummaryrefslogtreecommitdiff
path: root/config.yaml
diff options
context:
space:
mode:
authorPherkel2023-09-18 14:25:36 +0200
committerPherkel2023-09-18 14:25:36 +0200
commitd5e482b7dc3d8b6acc48a883ae9b53b354fa1715 (patch)
tree580f0ab45784664978d8f24c4831f3eec1bceb2e /config.yaml
parentd5689047fa7062b284d13271bda39013dcf6150f (diff)
decoder changes
Diffstat (limited to 'config.yaml')
-rw-r--r--config.yaml43
1 files changed, 26 insertions, 17 deletions
diff --git a/config.yaml b/config.yaml
index 41b473c..d248d43 100644
--- a/config.yaml
+++ b/config.yaml
@@ -1,3 +1,11 @@
+dataset:
+ download: True
+ dataset_root_path: "YOUR/PATH" # files will be downloaded into this dir
+ language_name: "mls_german_opus"
+ limited_supervision: False # set to True if you want to use limited supervision
+ dataset_percentage: 1.0 # percentage of dataset to use (1.0 = 100%)
+ shuffle: True
+
model:
n_cnn_layers: 3
n_rnn_layers: 5
@@ -6,32 +14,33 @@ model:
stride: 2
dropout: 0.3 # recommended to be around 0.4 for smaller datasets, 0.1 for really large datasets
+tokenizer:
+ tokenizer_path: "data/tokenizers/char_tokenizer_german.json"
+
+decoder:
+ type: "greedy" # greedy, or lm (beam search)
+
+ lm: # config for lm decoder
+ language_model_path: "data" # path where model and supplementary files are stored
+ language: "german"
+ n_gram: 3 # n-gram size of the language model, 3 or 5
+ beam_size: 50
+ beam_threshold: 50
+ n_best: 1
+ lm_weight: 2,
+ word_score: 0,
+
training:
- learning_rate: 5e-4
+ learning_rate: 0.0005
batch_size: 8 # recommended to maximum number that fits on the GPU (batch size of 32 fits on a 12GB GPU)
epochs: 3
eval_every_n: 3 # evaluate every n epochs
num_workers: 8 # number of workers for dataloader
-dataset:
- download: True
- dataset_root_path: "YOUR/PATH" # files will be downloaded into this dir
- language_name: "mls_german_opus"
- limited_supervision: False # set to True if you want to use limited supervision
- dataset_percentage: 1.0 # percentage of dataset to use (1.0 = 100%)
- shuffle: True
-
-tokenizer:
- tokenizer_path: "data/tokenizers/char_tokenizer_german.yaml"
-
-checkpoints:
+checkpoints: # use "~" to disable saving/loading
model_load_path: "YOUR/PATH" # path to load model from
model_save_path: "YOUR/PATH" # path to save model to
inference:
model_load_path: "YOUR/PATH" # path to load model from
- beam_width: 10 # beam width for beam search
- device: "cuda" # device to run inference on if gpu is available, else "cpu" will be set automatically
-lang_model:
- path: "data/mls_lm_german" #path where model and supplementary files are stored