-
Notifications
You must be signed in to change notification settings - Fork 256
Issues: InternLM/lmdeploy
[Benchmark] benchmarks on different cuda architecture with mo...
#815
opened Dec 11, 2023 by
lvhan028
Open
6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] Key Error loading OpenGVLab/Mini-InternVL-Chat-4B-V1-5
#1756
opened Jun 11, 2024 by
HaoLiuHust
2 tasks done
[Bug] Official image doesn't work for 4090 on CUDA 12.3 (but works for all other CUDA versions, and works for 12.3 on other GPU types)
#1750
opened Jun 11, 2024 by
josephrocca
2 tasks done
[Feature] Low priority: Allow specifying HuggingFace model/repo name in
lmdeploy convert
#1749
opened Jun 10, 2024 by
josephrocca
[Bug] Space is incorrectly removed from start of generated text for
/v1/completion
endpoint
#1743
opened Jun 8, 2024 by
josephrocca
2 tasks done
[Docs] Guidance on setting
num_tokens_per_iter
and max_prefill_iters
to optimal values
#1740
opened Jun 8, 2024 by
josephrocca
[Bug]
detokenize_incrementally
: OverflowError: out of range integral type conversion attempted
#1739
opened Jun 7, 2024 by
josephrocca
2 tasks done
High GPU memory for running InternVL-Chat-V1-5-AWQ
awaiting response
#1728
opened Jun 7, 2024 by
tairen99
[Bug] Mini-InternVL1.5-4B does not suceessfully initialized.
#1721
opened Jun 6, 2024 by
cydiachen
1 of 2 tasks
[Bug] internlm2-chat-1_8b模型使用4bit KV量化的时候找不到key_stats.pth
awaiting response
#1720
opened Jun 6, 2024 by
jxfruit
2 tasks
[Bug] Why does prefix caching change the generated content
#1719
opened Jun 5, 2024 by
DayDayupupupup
1 of 2 tasks
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.