forked from vllm-project/vllm
-
Notifications
You must be signed in to change notification settings - Fork 5
Insights: EmbeddedLLM/vllm
Overview
-
0 Active pull requests
-
- 0 Merged pull requests
- 0 Open pull requests
- 3 Closed issues
- 4 New issues
Loading
Could not load contribution data
Please try again later
Loading
3 Issues closed by 1 person
-
[Feature]: Try to add padding AITER Block Scaled Fused MoE of Qwen3
#53 closed
May 14, 2025 -
[Feature]: Add padding to ROCm AITER Block Scaled Fused MoE weights
#59 closed
May 14, 2025 -
[Feature]: Enhance the usage of rocm_aiter_fused_experts in fused_moe.py
#43 closed
May 13, 2025
4 Issues opened by 2 people
-
[Bug]: Fix the triton chunked prefill decode V1 bug
#63 opened
May 14, 2025 -
[Feature]: Integrate AITER MLA V1 from Upstream
#62 opened
May 14, 2025 -
[Feature]: MTP on ROCm
#61 opened
May 13, 2025 -
[Feature]: AITER Grouped TopK v1
#60 opened
May 13, 2025
2 Unresolved conversations
Sometimes conversations happen on old items that aren’t yet closed. Here is a list of all the Issues and Pull Requests with unresolved conversations.
-
[Feature]: Enable MLA for V1 on AMD [Triton MLA]
#46 commented on
May 13, 2025 • 0 new comments -
[Usage]: Check if lmcache works on ROCm
#56 commented on
May 13, 2025 • 0 new comments