-
Notifications
You must be signed in to change notification settings - Fork 4.2k
runs perfectly with the regular models, but not the quantized ones #993
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Ye same thing over here, sad hours |
I get the same error. I use Windows 10 with this version: whisper-bin-x64.zip
|
What happens if you build from source using latest |
Will there be a new official release to support the breaking change to the ggml models that occurred since April 30 by any chance? I get the same error as StaffanBetner with the 1.4.0 windows 64 exe release and downloading the q5 model from ggml.ggerganov so I assume this is the reason. Thanks for all your work on this project! Its amazing. |
I downloaded 1.4.2 from here: https://github.com/ggerganov/whisper.cpp/actions/runs/4973278607 and it works with the quantized models. Maybe it is a good idea to highlight in the readme that there are autocompiled versions under the action tab. |
Wow thanks, yes, I had no idea! appreciate you letting us know. |
Can reproduce this using CuBLAS on a RTX 3060 using v1.4.2. I don't want to use
|
Will probably make a new release after merging #1058 |
I have run quantization on ggml-small.en.bin to produce ggml-small.en.bin-q5_0.bin
Quantization proceeded without any errors.
When I run the model, it attempts to load, but throws the following:
GGML_ASSERT: ggml.c:4288: wtype != GGML_TYPE_COUNT
Abort trap: 6
Same issue occurs when I try 8-bit quantization.
Running on a 2015 Macbook Air, if this is any use.
The text was updated successfully, but these errors were encountered: