feat(deps): update dependency vllm-charts ( v0.14.1 → v0.15.0 ) #1663
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR contains the following updates:
v0.14.1→v0.15.0Warning
Some dependencies could not be looked up. Check the Dependency Dashboard for more information.
Release Notes
vllm-project/vllm (vllm-charts)
v0.15.0Compare Source
Highlights
This release features 335 commits from 158 contributors (39 new)!
Model Support
Engine Core
--async-schedulingnow works with pipeline parallelism (#32359).--enable-prefix-caching --mamba-cache-mode align. Achieves ~2x speedup by caching Mamba states directly (#30877).StreamingInputobjects while maintaining KV cache alignment (#28973).Hardware & Performance
NVIDIA
AMD ROCm
Other Platforms
Quantization
API & Frontend
include_stop_str_in_outputtuning (#32383),prompt_cache_keysupport (#32824).skip_special_tokensconfiguration (#32345).data_1/data_2andqueries/documents(#32577).avg_logprobandcompression_ratioin verbose_json segments (#31059).--ssl-ciphersCLI argument (#30937).api_server_countbased ondp_size(#32525), wheel variant auto-detection during install (#32948), custom profiler URI schemes (#32393).Dependencies
Breaking Changes & Deprecations
vllm:time_per_output_token_secondsmetric - usevllm:inter_token_latency_secondsinstead (#32661).Bug Fixes
New Contributors 🎉
Full Changelog: vllm-project/vllm@v0.14.1...v0.15.0
Configuration
📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
This PR has been generated by Renovate Bot.