Description
Name and Version
load_backend: loaded RPC backend from D:\139-conf\llama\llama\llama-b5844-bin-win-vulkan-x64\ggml-rpc.dll
ggml_vulkan: Found 1 Vulkan devices:
ggml_vulkan: 0 = AMD Radeon RX 7900 XT (AMD proprietary driver) | uma: 0 | fp16: 1 | warp size: 64 | shared memory: 32768 | int dot: 1 | matrix cores: KHR_coopmat
load_backend: loaded Vulkan backend from D:\139-conf\llama\llama\llama-b5844-bin-win-vulkan-x64\ggml-vulkan.dll
load_backend: loaded CPU backend from D:\139-conf\llama\llama\llama-b5844-bin-win-vulkan-x64\ggml-cpu-haswell.dll
version: 5844 (17a1f0d)
built with clang version 19.1.5 for x86_64-pc-windows-msvc
Operating systems
Windows
GGML backends
Vulkan
Hardware
5900x+7900xt
Models
qwen3 30b
Problem description & steps to reproduce
in 5688 version ,The chat can proceed normally.But in the latest version, conversations are limited to the first few hundred characters (300-500?)
Then it will output GGGGG at a slow speed (Approximately every ten seconds, a G is generated)
First Bad Commit
I think is 5808
I tested multiple versions and found that it worked fine under 5804, but then the issue mentioned above appeared