I have noticed that DeepSpeech 0.4.1 seems to perform particularly badly in situations where the user’s mouth is not directly in front of the microphone.
If you listen to the clips on the validation page, it seems that they are all performed by a user with a microphone right in front of their mouth.
This makes it harder to use for certain cases like a smart assistant that you talk to from across the other side of the room.
While some small improvement could perhaps be made by adding artificial reverb to the clips, are there plans to encourage more diverse recording environments in future?