-
Notifications
You must be signed in to change notification settings - Fork 410
Issues: InternLM/lmdeploy
[Benchmark] benchmarks on different cuda architecture with mo...
#815
opened Dec 11, 2023 by
lvhan028
Open
9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] pytorch backend 's precision points loss 1.0-2.5 points between main code and v0.6.1 on some models.
#2679
opened Oct 29, 2024 by
zhulinJulia24
3 tasks
[Bug] no user input makes api server throw exception with MLLM
#2658
opened Oct 25, 2024 by
gaord
3 tasks
[Bug] use new 4bits quantizated models of internlm2, decoded word starts with a blank.
#2651
opened Oct 24, 2024 by
zhulinJulia24
3 tasks
[Bug] Use triton to deploy minicpm-v-2_6 GPU memory keeps increasing until it overflows
#2642
opened Oct 24, 2024 by
LinJianping
1 of 3 tasks
[Bug] Phi-3-vision-128k-instruct 跑模型在8卡上出现 “Expected all tensors to be on the same device, but found at least two devices”
mllm
#2633
opened Oct 22, 2024 by
dreamerlin
3 tasks done
[Bug] qwen2-vl-7b docker delpoy bugs
awaiting response
#2629
opened Oct 22, 2024 by
jnzbfgjd
3 tasks done
[Feature] Combine Batched Inference and Chat Conversation in VLMs Deployment
#2628
opened Oct 21, 2024 by
Yusepp
[Bug] When TP = 4 and prefix cache is enabled, no result is generated.
#2611
opened Oct 17, 2024 by
rbao2018
3 tasks done
[Bug] InternVL2-26B model load extremely slow
#2608
opened Oct 16, 2024 by
HappyNotHappy
3 tasks done
Previous Next
ProTip!
Find all open issues with in progress development work with linked:pr.