I’ve been running different fine tuning tests on the 0.1.1 release by just training on single voices to see if I could improve the model performance, I train with the default command:
python3 DeepSpeech.py --n_hidden 2048 --initialize_from_frozen_model path/to/model/output_graph.pb --checkpoint_dir fine_tuning_checkpoints --epoch 3 --train_files my-train.csv --dev_files my-dev.csv --test_files my_dev.csv --learning_rate 0.0001
I trained with 1 hour of audio for 1-3 epochs and the model never seems to improve. I’m wondering if I’m doing something wrong or the code is not working. Has anyone else here successfully trained a better performing model using fine tuning on top of deepspeech?