You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Some CPUs such as ARM Neoverse-N1 (Oracle Cloud free tier) support FP16 computation.
It would be nice to have this feature because there could be up to a 2x speedup in computation speed compared to float32.
I'm just creating the enhancement request and understand that there might not be a focus on this because it only applies to a small subset of CPUs.
The text was updated successfully, but these errors were encountered:
I would love to use faster-whisper instead of whisper.cpp but the lack of FP16 on CPU is kind of the deal breaker for me.
For now, with faster-whisper I have to choose between speed (int8) and accuracy (float32), FP16 is the missing balance point.
Consider that ARM CPUs are very popular nowadays (mobile phone, Apple M series, cloud providers)
I think this is a good idea to consider on fp16 support.
Do you have any thoughts about this @guillaumekln
Not just arm and apple silicon, latest intel and AMD CPUs (I am using an AMD one) also support float16. Would be very nice if int8_int16 inference is supported on the CPUs which support float16.
From SYSTRAN/faster-whisper#65
Some CPUs such as ARM Neoverse-N1 (Oracle Cloud free tier) support FP16 computation.
It would be nice to have this feature because there could be up to a 2x speedup in computation speed compared to float32.
I'm just creating the enhancement request and understand that there might not be a focus on this because it only applies to a small subset of CPUs.
The text was updated successfully, but these errors were encountered: