You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Im trying to build an Arabic asr with common voice dataset with pretrained stt_en_citrinet_256 and after a 120 epochs i found that my model didn't improve and still making a bad WER (0.95) on test set.
i used sub-word sentencepiece tokenizer with 4096 size and changed the pretrained model vocab, tokenizer, train and test config as below:
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
Hello,
Im trying to build an Arabic asr with common voice dataset with pretrained stt_en_citrinet_256 and after a 120 epochs i found that my model didn't improve and still making a bad WER (0.95) on test set.
i used sub-word sentencepiece tokenizer with 4096 size and changed the pretrained model vocab, tokenizer, train and test config as below:
from omegaconf import OmegaConf, open_dict
params = OmegaConf.load("./configs/config_bpe.yaml")
params.model.train_ds.manifest_filepath = 'train.json'
params.model.validation_ds.manifest_filepath = 'dev.json'
first_asr_model.change_vocabulary( new_tokenizer_dir='tokenizer_spe_unigram_v4096', new_tokenizer_type="bpe" )
first_asr_model.setup_training_data(train_data_config=params['model']['train_ds'])
first_asr_model.setup_validation_data(val_data_config=params['model']['validation_ds'])
below is my train, val loss and WER
Beta Was this translation helpful? Give feedback.
All reactions