Replies: 4 comments 19 replies
-
|
If you do speed benchmarks then use Try "256.1" [only Pro is released at the moment] Similar thread: #442 |
Beta Was this translation helpful? Give feedback.
-
|
@subgrinder: Good catch, but the typo --sentence has been luckily only in my post, not in my command line. Maybe a stupid question, but I hadn't yet the time to read all the threads about "batched". Is the quality of batched affected by the number or ist only lower if turned on? Does batch_size 32 result in lower quality than 8? Does the batched mode also affect the quality and speed of spiaker_diarization? I could not tested it yet, because the standard version runs on error on my 5090. I have done several test runs again. The results are:
|
Beta Was this translation helpful? Give feedback.
-
|
The benchmark script has been expanded. I tested two files: 12min and 40min with Pro r3.256.1 on RTX5090. RTF = Real Time Factor (higher is faster) RTX 5090 video duration 12 min RTX 5090 video duration 40 min @Purfview It's late now, I will test the three patches tomorrow and adjust/shorten the benchmark script. |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.














Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello Mr. Purfview.
I replaced my old RTX 6000 (2018) with an actual RTX 5090 (2025). I excpeted 3times faster results, because of the FP16 values (33 vs 105 TFLOPS). But the RTX5090 is not faster, in some cases is it even a little bit slower. I also updated the nvidia driver to latest v581.57 on my Workstation Dell Precision 5820 with CPU intel core i9-10900x @3.7GHz.
I did several test runs with faster-whisper-xxl 245.4 on Win11 25H2. I tried --compute_type float16 and --compute_type float32 and decided for float32, because it computes faster than 16. (I got the cuBLAS failed error message without --compute_type)
Results:

Example:

.\faster-whisper-xxl.exe BAM.mp4 --model large-v2 --language de --compute_type float32Why is the transcribe process not faster, is it a software options issue?
Any advice is appriciated.
Beta Was this translation helpful? Give feedback.
All reactions