Skip to content

Issues: vllm-project/vllm

[Roadmap] vLLM Roadmap Q4 2024
#9006 opened Oct 1, 2024 by simon-mo
Open 26
vLLM's V1 Engine Architecture
#8779 opened Sep 24, 2024 by simon-mo
Open 9
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

[Bug]: base64 string leads to gibberish with latest vLLM server and pixtral-12b bug Something isn't working
#11781 opened Jan 6, 2025 by michael-brunzel
1 task done
[Bug]: prompt logprobs are different with batch_size > 1 compared to batch_size=1 bug Something isn't working
#11778 opened Jan 6, 2025 by rizar
1 task done
[Bug]: bug Something isn't working
#11775 opened Jan 6, 2025 by kzos
1 task done
[Usage]: Running OpenAI Swarm with vLLM-hosted models usage How to use vllm
#11774 opened Jan 6, 2025 by ArturDev42
1 task done
[Bug]: Cutlass 2:4 Sparsity + FP8/Int8 Quant RuntimeError: Error Internal bug Something isn't working
#11763 opened Jan 6, 2025 by leoyuppieqnew
1 task done
[Misc]: Very High GPU RX/TX using vllm misc
#11760 opened Jan 6, 2025 by alexpong0630
1 task done
[Doc]: Why NGramWorker does not support cache operations documentation Improvements or additions to documentation
#11758 opened Jan 6, 2025 by kuangdao
1 task done
[Bug]: CPU Offload fails when enable_lora=True bug Something isn't working
#11748 opened Jan 5, 2025 by Neko-nos
1 task done
[Performance]: Context Length Problem with VLLM performance Performance-related issues
#11745 opened Jan 5, 2025 by MotorBottle
1 task done
[Installation]: XPU dependencies not built against most recent oneAPI installation Installation problems
#11734 opened Jan 4, 2025 by janimo
1 task done
[Usage]: serving 'LLaVA-Next-Video-7B-Qwen2' usage How to use vllm
#11731 opened Jan 4, 2025 by Noctis-SC
1 task done
[Bug]: PixtralHF inference broken since #11396 bug Something isn't working
#11726 opened Jan 3, 2025 by mgoin
1 task done
[New Model]: unsloth/Llama-3.3-70B-Instruct-bnb-4bit new model Requests to new models
#11725 opened Jan 3, 2025 by Hyfred
1 task done
[Bug]: 0.6.6.post1 crash in marlin_utils.py bug Something isn't working
#11703 opened Jan 3, 2025 by Flynn-Zh
1 task done
[Bug]: vLLM LoRA Crash when using Dynamic Loading bug Something isn't working
#11702 opened Jan 3, 2025 by haitwang-cloud
1 task done
ProTip! Adding no:label will show everything without a label.