Skip to content

Issues: NVIDIA/TensorRT-LLM

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

The accuracy of trt-llm-qwen-vl-chat is low. bug Something isn't working
#2241 opened Sep 19, 2024 by xiangxinhello
2 of 4 tasks
Cannot set earlyStopping 0 when using ModelRunnerCpp bug Something isn't working
#2239 opened Sep 18, 2024 by PKaralupov
4 tasks
How to set TensorRT-LLM to use Flash Attention 3 question Further information is requested triaged Issue has been triaged by maintainers
#2238 opened Sep 18, 2024 by kanebay
Working with vllm is much easier than working with tensorrt feature request New feature or request triaged Issue has been triaged by maintainers
#2237 opened Sep 18, 2024 by Alireza3242
gemma-2-27b bad outputs bug Something isn't working
#2233 opened Sep 17, 2024 by siddhatiwari
2 of 4 tasks
FP8 rowwise support possible for SM89?
#2229 opened Sep 17, 2024 by aikitoria
why do ouput include <|im_end|>
#2228 opened Sep 14, 2024 by w066650
whisper-medium decoder Compile blocking bug Something isn't working
#2227 opened Sep 14, 2024 by skyCreateXian
1 of 4 tasks
"use_embedding_sharing" option not working for llama model. bug Something isn't working
#2226 opened Sep 14, 2024 by jxchenus
2 of 4 tasks
LLM configured host_cache_size is invalid
#2225 opened Sep 14, 2024 by ljayx
error in tag v0.12.0 when build from source bug Something isn't working
#2222 opened Sep 12, 2024 by zhangts20
2 of 4 tasks
Support for florence-2
#2221 opened Sep 11, 2024 by Source82
nccl ops from TRT-LLM
#2220 opened Sep 11, 2024 by apbose
Cannot build quantized int8 models for Phi3 128k models [TensorRT-LLM 0.12.0] bug Something isn't working
#2214 opened Sep 10, 2024 by louis845
2 of 4 tasks
multi-gpu error:MPI_Unknown_error for examples/apps/chat.py bug Something isn't working
#2209 opened Sep 9, 2024 by youxzAnt
4 tasks
Qwen-VL-Chat has an error bug Something isn't working
#2206 opened Sep 9, 2024 by xiangxinhello
2 of 4 tasks
ProTip! Type g p on any issue or pull request to go back to the pull request listing page.