-
Notifications
You must be signed in to change notification settings - Fork 162
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to reduce Whisper hallucinations #198
Comments
Something is missing in this description. |
@Jeronymous this is the code I called:
(changing thetemperature to
skipping the segment and word level timestamps as I am not sure they are relevant for this situation. |
OK it seems the model hallucinates completely. Otherwise, I would need the audio to investigate more. Maybe the volume level is particularly low ? You can also try option vad="auditok" or vad="silero", to try to avoid hallucination, as well as "condition_on_previous_text=False" |
Thank you @Jeronymous changing the temperature and adding vad="auditok", did indeed help at least it shows correct text, just generates loads of duplications still. I am sceptical on adding the |
When I ran 15min audio through whisper timestamped model "", which is from an interview - it randomly shows completely unrelated text in the audio that was played. Has anyone ever see such an issue?
The text was updated successfully, but these errors were encountered: