So if one were to use DeepSpeech for live transcription (like Google’s Live Transcribe, for instance), they would need to implement a punctuation model as well.
Has anyone looked into how this might work? Or are there already Python libraries that do this? So you would run your DeepSpeech output through those models and hopefully get well-punctuated and more readable sentences.
I know that this issue may be out of scope for the DeepSpeech project, but is relevant to the overall context in which a DeepSpeech solution might be used. So, thoughts?