-
Notifications
You must be signed in to change notification settings - Fork 380
Issues: InternLM/lmdeploy
[Benchmark] benchmarks on different cuda architecture with mo...
#815
opened Dec 11, 2023 by
lvhan028
Open
9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug]TypeError: TextEncodeInput must be Union[TextInputSequence, Tuple[InputSequence, InputSequence]]
#2476
opened Sep 18, 2024 by
LIUKAI0815
3 tasks
[Bug] The parameters
n=number_outputs
does not work in the function: response = client.chat.completions.create
awaiting response
#2474
opened Sep 17, 2024 by
leoozy
3 tasks done
[Bug] 2x4090 with Llama2 70B silently crashes (i.e. without any error message in DEBUG mode) as of v0.6.0a0 and v0.6.0 (but works fine in previous versions)
#2468
opened Sep 14, 2024 by
josephrocca
3 tasks done
[Bug] output not consistent with different max_prefill_token_num for long context input on pytorch engine
#2457
opened Sep 12, 2024 by
RunningLeon
3 tasks done
[Bug] CUDA runtime error when running Llama-3.1-70B-Instruct-AWQ-INT4
awaiting response
#2442
opened Sep 10, 2024 by
rtadewald
1 of 3 tasks
[Bug] lmdeploy does not support the regularized lora target module
awaiting response
#2439
opened Sep 10, 2024 by
orzgugu
2 of 3 tasks
[Bug] 请问一下,我用qwen2的模型,仿照internlm-xcomposer2中的plora方式训练了一个VL模型,这个模型应该怎样用lmdeploy部署起来
awaiting response
#2437
opened Sep 10, 2024 by
alanayu
3 tasks
[Bug] 部署internvl2自己lora微调的模型,显存占用非常高,快速爬升到了60g正常吗
#2425
opened Sep 5, 2024 by
yywangfei
1 of 3 tasks
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.