Replies: 2 comments 1 reply
-
|
Read there -> Some problems with large-v3
That's normal.
Yes, test with same r160.3 version. |
Beta Was this translation helpful? Give feedback.
1 reply
-
|
Looks like a general consensus is that Whisper-v3 Hallucinations on Real World Data Whisper's original large-v3 tread: openai/whisper#1762 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I needed to create subtitles for a 90 min documentary so thought it would be a decent test to see if the new v3-large model had any significant improvements over the v2-large.
This post is not about an issue though, just my findings.
I upgraded to
Whisper-Faster_r160.3_windowsand then ran these commands against the same video file.whisper-faster "%vidname%" -l=en -m=large-v2 -ct=int8_float32 -o=source --verbose=true -bs=1using the v2-large model
and
whisper-faster "%vidname%" -l=en -m=large-v3 -ct=int8_float32 -o=source --verbose=true -bs=1using the v3-large model
I compared the two outputs in Subtitle Edit and there were plenty of differences, too many to list here.
Is this a fair comparison or should I have used the old binary (r160) with the v2-large model and then the new binary with the v3-large model?
Just remembered I can change the verbose command to just -v
Beta Was this translation helpful? Give feedback.
All reactions