forked from vllm-project/vllm
-
Notifications
You must be signed in to change notification settings - Fork 134
Pull requests: HabanaAI/vllm-fork
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Delay prefix cache calculation to find longest common prefix
#2170
opened Dec 8, 2025 by
ikurtchen
Loading…
3 tasks done
Bump actions/stale from 9.1.0 to 10.1.1
dependencies
Pull requests that update a dependency file
github_actions
Pull requests that update GitHub Actions code
#2169
opened Dec 8, 2025 by
dependabot
bot
Loading…
[DeepSeek R1] Gracefully shutdown mooncake store when exception for kill
#2166
opened Dec 4, 2025 by
jerrychenhf
Loading…
Enable delayed sampling for warmup also to remove graph compilation i…
#2164
opened Dec 4, 2025 by
yeonsily
Loading…
3 tasks
add VLLM_ENGINE_PROFILER_SKIP_STEPS to the engine profiler
#2143
opened Nov 19, 2025 by
yangulei
Loading…
[DeepSeek R1] chunked prefill warmup with chunk size
#2135
opened Nov 14, 2025 by
jerrychenhf
Loading…
Workaround for Assertion error when embedding with bge-m3 in lazy mode
#2093
opened Oct 28, 2025 by
slokesha
Loading…
fix bug that VLLM_SKIP_WARMUP=1 is not recognized in vision_bucket
#2036
opened Oct 15, 2025 by
yingjie-han
Loading…
Previous Next
ProTip!
Filter pull requests by the default branch with base:habana_main.