-
-
Notifications
You must be signed in to change notification settings - Fork 5.9k
Issues: vllm-project/vllm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug]: mllama AssertionError during kv cache profiling
bug
Something isn't working
#13929
opened Feb 26, 2025 by
tjohnson31415
1 task done
[Usage]: Qwen2-VL-2B-Instruct Issue when passing a video URL to /chat/completions
usage
How to use vllm
#13927
opened Feb 26, 2025 by
cquil11
1 task done
[Bug]: always finish_reason='length' using google/gemma-2-27b-it
bug
Something isn't working
#13924
opened Feb 26, 2025 by
yananchen1989
1 task done
[Bug]: Speculative Decoding Tokens not being included in Prometheus metrics
bug
Something isn't working
#13916
opened Feb 26, 2025 by
nimapourjafar
[Bug]: running deepseek-r1 14B with 2*5090D
bug
Something isn't working
#13914
opened Feb 26, 2025 by
lambo111-x86
1 task done
[New Model]: Magma Model Support
new model
Requests to new models
#13913
opened Feb 26, 2025 by
Praveenstein
1 task done
[Feature]: support binding on Unix Domain Sockets (UDS)
feature request
#13907
opened Feb 26, 2025 by
neoq
1 task done
[Feature]: T5Model has no vLLM implementation
feature request
#13903
opened Feb 26, 2025 by
dromeuf
1 task done
[Bug]: 【Qwen2.5-VL-72B-Instruct-AWQ】ERROR 02-26 05:28:06 engine.py:400] Error while deserializing header: InvalidHeaderDeserialization
bug
Something isn't working
#13899
opened Feb 26, 2025 by
jieguolove
1 task done
[Feature]: Upstream flash attention to support cutlass 3.8
feature request
#13893
opened Feb 26, 2025 by
johnnynunez
1 task done
[Bug]: ValueError: Unsupported config format: ConfigFormat.AUTO on macOS
bug
Something isn't working
#13889
opened Feb 26, 2025 by
unizhu
1 task done
[Bug]: collect_env doesn't work in uv environment
bug
Something isn't working
#13888
opened Feb 26, 2025 by
bmuskalla
1 task done
[Bug]: vllm部署qwen2.5_vl_72b之后,你们有出现,刚部署好之后调用一切正常3-5秒一条,然后使用一段时间,就越来越慢了的情况吗60s一条
bug
Something isn't working
#13886
opened Feb 26, 2025 by
lmingze
1 task done
[Feature]: Any plan run deepseek-r1 fp8 on Ampere gpu
feature request
#13885
opened Feb 26, 2025 by
Arcmoon-Hu
1 task done
[Bug]: vllm 0.7.3, system gets stuck during the reasoning process
bug
Something isn't working
#13884
opened Feb 26, 2025 by
bigMam
1 task done
[Bug]: Incorrect first_token_time and first_scheduled_time metrics results
bug
Something isn't working
#13883
opened Feb 26, 2025 by
iboiko-habana
1 task done
[Usage]: I want to be able to Qwen2.5-7B & RTX4060
usage
How to use vllm
#13882
opened Feb 26, 2025 by
sancelot
1 task done
[Bug]: vllm v0.7.3 - The following fields were present in the request but ignored: {'top_logprobs'}
bug
Something isn't working
#13881
opened Feb 26, 2025 by
tonyaw
1 task done
[Bug]: top_logrpobs generating a WARNING
bug
Something isn't working
#13880
opened Feb 26, 2025 by
FilippoBoni1921
1 task done
[Feature]: Support Deepseek's DeepGemm MoE
feature request
#13879
opened Feb 26, 2025 by
gzy19990617
1 task done
[New Model]: dunsloth/DeepSeek-R1-GGUF
new model
Requests to new models
#13877
opened Feb 26, 2025 by
tpoc
1 task done
[Usage]: How to use vllm
max_num_batched_tokens
and max_model_len
usage
#13875
opened Feb 26, 2025 by
tingjun-cs
1 task done
[Usage]: when i deploy a model ,how to set the max input str length and the number of max input token. and the max output length??
usage
How to use vllm
#13874
opened Feb 26, 2025 by
lmingze
1 task done
[Bug]: V1 does not support torch compile
bug
Something isn't working
#13872
opened Feb 26, 2025 by
echozyr2001
1 task done
[Feature]: Support microsoft/Magma-8B
feature request
#13870
opened Feb 26, 2025 by
yxchng
1 task done
Previous Next
ProTip!
Adding no:label will show everything without a label.