diff options
author | Pherkel | 2023-09-18 14:25:36 +0200 |
---|---|---|
committer | Pherkel | 2023-09-18 14:25:36 +0200 |
commit | d5e482b7dc3d8b6acc48a883ae9b53b354fa1715 (patch) | |
tree | 580f0ab45784664978d8f24c4831f3eec1bceb2e /config.yaml | |
parent | d5689047fa7062b284d13271bda39013dcf6150f (diff) |
decoder changes
Diffstat (limited to 'config.yaml')
-rw-r--r-- | config.yaml | 43 |
1 files changed, 26 insertions, 17 deletions
diff --git a/config.yaml b/config.yaml index 41b473c..d248d43 100644 --- a/config.yaml +++ b/config.yaml @@ -1,3 +1,11 @@ +dataset: + download: True + dataset_root_path: "YOUR/PATH" # files will be downloaded into this dir + language_name: "mls_german_opus" + limited_supervision: False # set to True if you want to use limited supervision + dataset_percentage: 1.0 # percentage of dataset to use (1.0 = 100%) + shuffle: True + model: n_cnn_layers: 3 n_rnn_layers: 5 @@ -6,32 +14,33 @@ model: stride: 2 dropout: 0.3 # recommended to be around 0.4 for smaller datasets, 0.1 for really large datasets +tokenizer: + tokenizer_path: "data/tokenizers/char_tokenizer_german.json" + +decoder: + type: "greedy" # greedy, or lm (beam search) + + lm: # config for lm decoder + language_model_path: "data" # path where model and supplementary files are stored + language: "german" + n_gram: 3 # n-gram size of the language model, 3 or 5 + beam_size: 50 + beam_threshold: 50 + n_best: 1 + lm_weight: 2, + word_score: 0, + training: - learning_rate: 5e-4 + learning_rate: 0.0005 batch_size: 8 # recommended to maximum number that fits on the GPU (batch size of 32 fits on a 12GB GPU) epochs: 3 eval_every_n: 3 # evaluate every n epochs num_workers: 8 # number of workers for dataloader -dataset: - download: True - dataset_root_path: "YOUR/PATH" # files will be downloaded into this dir - language_name: "mls_german_opus" - limited_supervision: False # set to True if you want to use limited supervision - dataset_percentage: 1.0 # percentage of dataset to use (1.0 = 100%) - shuffle: True - -tokenizer: - tokenizer_path: "data/tokenizers/char_tokenizer_german.yaml" - -checkpoints: +checkpoints: # use "~" to disable saving/loading model_load_path: "YOUR/PATH" # path to load model from model_save_path: "YOUR/PATH" # path to save model to inference: model_load_path: "YOUR/PATH" # path to load model from - beam_width: 10 # beam width for beam search - device: "cuda" # device to run inference on if gpu is available, else "cpu" will be set automatically -lang_model: - path: "data/mls_lm_german" #path where model and supplementary files are stored |