-
Notifications
You must be signed in to change notification settings - Fork 11.8k
Misc. bug: llama-sampling.cpp:204: GGML_ASSERT(cur_p->size > 0) failed #13405
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
...I should have added a --seed flag, but the issue is reproducible for me with all seeds I've tried so far. The issue has to do with
If I adjust this to
|
I'm seeing this bug as well, I'm not passing in --dry-allowed-length 4. main: server is listening on http://0.0.0.0:8089 - starting the main loop |
I can confirm the same behavior on MacOS. Error: que start_loop: waiting for new tasks command line to start llama-server: /Users/myuserdir/Projects/llamacpp/bin/llama-server JSON payload: request: { |
Name and Version
Operating systems
Linux
Which llama.cpp modules do you know to be affected?
llama-cli
Command line
Problem description & steps to reproduce
The log file of the output, together with what I hope is all the relevant information can be found in this ephemeral repo I put up for this bug report:
https://github.com/bjodah/bug-reproducer-llamacpp-assert-triggering/tree/main
It might very well that I'm doing something awfully wrong here, but since it's an assert that is triggering, I'm thinking that you might be interested in a bug report?
I first observed this error using llama-serve on my laptop (ubuntu 24.04, geforce 1050 mobile), but everything in this bug report was reproduced on a more modern system (debian, geforce rtx 3090).
First Bad Commit
Qwen 3 support is pretty recent, so I haven't figured out what's the relevant oldest commit for a bisection.
Relevant log output
The text was updated successfully, but these errors were encountered: